Thank you community for translating an earlier article into Portuguese.
Am returning the favor with a new release of Pattern Match Workbench demo app.
Added support for Portuguese.
The labels, buttons, feedback messages and help-text for user interface are updated.
Pattern Descriptions can be requested for the new language.

The single AI Model for transforming user prompt into Pattern match code was fully retrained.
.png)
Values to Pattern Code Model also retrained
The separate AI model for generating Pattern match code from a sample list of values has been retrained.
The previous model release exhibited loss of generalized pattern identification behaviors in order to support new delimiter style patterns.
The new training approach was evolved to better suit both types of challenge.
Post training benchmarking of the model stages was used to quantify the better candidate to preserve generalized pattern solving.
The principle to quantify 100% success for each generated pattern was to attempt to satisfy all its respective candidate sample records.
Overview benchmark report:
The following table gives examples from benchmark candidates demonstrating partial success.
It shows the percentage of sample records successfully matched to generated pattern code.
Legend
Sample Size - The number records available for test sample. Sometimes a pattern describes less than 31 possible exact matches. Hence the lower values
Context Window - Maximum number of sample records that fit to generate a new pattern
Rows matched - This is the number of sample records that were matched by the generated pattern
Note that items 4, 6 and 8 achieve a match count greater than the context window.
When patterns are generated from values the context window used and proportion of matching is returned.
For challenging data samples it can be productive to retry the "Pattern from Values" button for improved results.
Showing the context window by row is useful to provide opportunity for specific records to be elevated into prompt for generative consideration.
.png)
Benchmark inferencing was conducted via HuggingFace spaces to scale up available GPU for timely execution.
Reporting is useful to analyse qualitative closeness of proposals by the LLM in terms of:
* Match type ( String, Number, Alphanumeric, Uppercase-Alpha)
* Choice of Quantity range ( 2-to-3 of ... )
* Option lists ( "123" or "456" )
* Structured elements using brackets
This opens up opportunities for training to be improved.
Please feel free to feedback on improvements for:
- Language translations
- Pattern match behaviors.