Replies: 2 comments 3 replies
-
|
We should definitely investigate whether Transformers can be applied here. You are correct that we are focused on the performance of inference, not training. As far as GPUs, the model needs to be able to be run on-device on all types of hardware. This means everything from high-resource data centers to low-resource smartphones. Even on devices with a GPU available, we should be cognizant of battery use. A project primarily focused on adding CUDA support wouldn't be compelling; we already know we can do that. We'd like to make the model itself more efficient, ideally to the point that it doesn't need CUDA. The baseline is that we have a hand-crafted trie-based dictionary data structure that is very fast but is not as accurate and uses more data. The LSTM is smaller and more accurate but uses more processing power. Any new model should be benchmarked against both the dictionary and the LSTM, in environments without GPU. |
Beta Was this translation helpful? Give feedback.
-
|
Hi @sffc I have submitted my proposal for “Using AI to Better Segment Complex Scripts” I really hope it doesn’t look like total LLM spam 🙇♀️—I put a lot of effort into organizing it clearly. Thanks again for all your time and guidance! I’ve learned a lot from the Unicode developers, and I’m really happy to have been part of the process—even if I was not selected. |
Beta Was this translation helpful? Give feedback.

Uh oh!
There was an error while loading. Please reload this page.
-
I’ve realized that this segmentation system was originally designed around 2019–2020, at a time when Transformer-based architectures were still emerging and NVIDIA GPUs were primarily seen as gaming hardware rather than standard ML infrastructure. The system reflects that era: it relies on LSTM and AdaBoost models, and I found no evidence of built-in GPU acceleration in the current codebase.
Given that, I believe we should first modernize the current baseline by:
Although I understand that the performance of segmentation here mostly refers to inference rather than training, optimizing GPU utilization during large-scale evaluation, multilingual experiments, or re-training could significantly accelerate development iterations and model tuning.
Adding optional GPU support could bring:
If my GSoC proposal would be accepted, I’d like to submit a pull request to modularize GPU/CPU device handling in
pick_lstm_model()and benchmark the speed and memory performance of both modes.Would that be necessary?
By the way, I would like to sincerely thank the mentors at Unicode for their patience and support. I know I've asked quite a few basic or naive questions along the way, and I truly appreciate the time and insight you've shared.I’ve been wandering through different parts of the Unicode community of GSoC, and this project feels like the best fit for my technical stack so far.
Beta Was this translation helpful? Give feedback.
All reactions