Most likely. Our middleware integrated features realizing high efficiency based on the latest research papers. However, the speedup factor depends on your applications. Please consider our GPU workload analysis service for more details.
No. In general, your script is runnable since Fixstars AI Booster is based on de-facto standard GenAI/LLM OSS middleware. Additional changes like adding arguments to your script would accelerate training/inference performance.
Yes. Our engineers regularly follow up latest research papers/OSS and port them to Fixstars AI Booster. You can enjoy state of the art LLM technologies. Please contact us if you would like to know which technologies are supported.
Yes. Some basic parts are common in GenAI/LLM and regular DNN. Thus, you can boost your AI/ML applications with Fixstars AI Booster. Please contact us for further details.
We optimized LLM middleware to the target servers. For instance, we tuned multi-GPU/multi-node communication and file storage for our cloud environment, resulting in 2.1x-3.7x performance boots.
Fixstars engineers acquired knowledge and skills to accelerate a large variety of computations including AI/ML from our engineering services. So we can apply acceleration techniques we have cultivated so far to this field.
For LLM pre-training, you can accommodate up to 32B models. For LLM fine-tuning, you can accommodate up to 70B models. For further questions, please contact us. Our experts will support you.
Preview releases will be per-month basis and regular official updates will be quarterly-basis.