
Fixstars
@Fixstars_US
Followers
190
Following
14
Media
31
Statuses
134
Innovators in Software Acceleration Solutions.
Irvine, CA
Joined June 2011
π΄ Fixstars is exhibiting at the Generative AI Summit Los Angeles!.π
August 27, 2025.π The LINE LA.π Event details: Stop by to explore how Fixstars AI Booster accelerates generative AI workloads with cutting-edge performance engineering. #Fixstars.
world.aiacceleratorinstitute.com
Join Generative AI Summit Los Angeles alongside hundreds of pioneering engineers, developers & executives that are facilitating the latest tech revolution.
0
1
2
π Accelerate Your AI Workloads: Performance Engineering with Fixstars AIBooster.π [Register Now] Discover how to maximize AI infrastructure efficiency and significantly reduce operational costs with hands-on performance engineering strategies. Join our.
fixstars.com
Featuring real-world examples and case studies using Fixstars AIBooster, attendees will learn GPU profiling techniques, job-based performance tracking, and effective methods for identifying and...
0
0
1
Boost your ROS 2 data storage efficiency! Our new blog introduces rosbag2_storage_mcap, integrating MCAP for better performance & easier recovery. See file size comparisons with SQLite3 and why MCAP's set to be the ROS 2 default!.Learn more:
blog.us.fixstars.com
The rosbag2_storage_mcap package has been added, allowing MCAP to be selected as the storage for rosbag2. This article introduces how to use rosbag2_storage_mcap.
0
0
1
Llama 4 Scout on a single H100 with INT4? Our new blog confirms it loads, but reveals surprising context limits. Crucial insights for your LLM deployments. Read more:
blog.us.fixstars.com
We verified the lightweighting of Llama 4 Scout, announced by Meta, through INT4 quantization using the official implementation. We will report in detail on its execution on a single H100 and the...
0
0
1
Fine-tuning Llama 4 Scout? Our new blog dives into performance engineering with LLaMA-Factory! Learn about setup, LoRA/DeepSpeed, and optimizing GPU utilization for faster training. Master LLM fine-tuning: #Llama4 #FineTuning #PerformanceEngineering.
blog.us.fixstars.com
Fine-tuning Llama 4 Scout using LLaMA-Factory and DeepSpeed, and implementing speedups and GPU optimization through batch size adjustments. We will explain the procedure in detail.
0
1
2
Unlock the full potential of Llama 4 with vLLM: context length matters!.Our new blog details optimal settings for Llama 4 Scout & Maverick models, with GPU configs & quantization insights. Optimize your LLM deployments: #Llama4 #vLLM.
blog.us.fixstars.com
Llama 4 supports a long context of up to 10 million tokens; however, for actual deployment, it's important to be mindful of library and hardware constraints. In this article, through experiments...
0
0
0
Gemma 3 (27B) locally? Our new blog post dives into deploying this powerful LLM, which has shown top-tier performance even against closed models! We cover setup with Ollama and vLLM for personal & team environments. Unlock Gemma 3's potential: #Gemma3.
blog.us.fixstars.com
In this article, we actually implemented the much-talked-about 'Gemma 3 27B model' using ollama and vllm, and conducted performance verification on specific work-related tasks such as code understa...
0
2
2
π Experience AI performance observability β no installation needed. π Try the live demo now: With Fixstars AI Booster, you can apply performance engineering to your AI workloads:.βοΈ Optimize models.πΈ Cut GPU costs.π Gain real-time insights. #AI
0
0
0
Want to run Llama 4 Scout cost-effectively? Our blog shows you how to leverage RTX 6000 Ada GPUs with llama.cpp as a more accessible alternative to the pricey H100. See how: #llama_cpp #RTX6000Ada #TechTips.
blog.us.fixstars.com
In a previous verification, we used a server equipped with an NVIDIA H100 GPU to run Llama 4 Scout. The H100 is expensive, and its implementation locations are limited due to power consumption and...
0
0
1