⚙️ System Requirements
- 🧠 Memory: 32 GB RAM or higher recommended
- ⚡ CPU/NPU: AMD Ryzen™ AI laptop with XDNA2 NPU
- 🖥️ OS: Windows 11
While FastFlowLM can run with 16 GB RAM, complex models (e.g., 3B or 8B) may require >= 32 GB for optimal performance and longer context length (more kv cache).
🚨 CRITICAL: NPU Driver Requirement
You must have AMD NPU driver version number >= 32.0.203.311 installed for FastFlowLM to work correctly.
- Check via:
Task Manager → Performance → NPU
or
Device Manager → NPU
💾 Installation (Windows)
A packaged FLM Windows installer is available here:
flm-setup.exe
To install silently, open PowerShell, navigate (
cd) to the download folder, and run:
.\flm-setup.exe /Silent
# or
.\flm-setup.exe /VERYSilent
If you see “Windows protected your PC”, click More info, then select Run anyway.
For version history and changelog, see the release notes.
🚀 NPU Power Mode
By default, FLM runs in performance NPU power mode. You can switch to other NPU power modes (powersaver, balanced, or turbo) using the --pmode flag:
CLI mode:
flm run gemma3:4b --pmode balanced
Server mode:
flm serve gemma3:4b --pmode balanced
⚠️ Note: Using powersaver or balanced will lower NPU clock speeds and cause a significant drop in speed. For more details about NPU power mode, refer to the AMD XRT SMI Documentation.
🧪 Quick Test (CLI Mode)
After installation, do a quick test to see if FastFlowLM is properly installed. Open PowerShell (Win + X → I), and run a model in terminal (CLI mode):
flm run llama3.2:1b
Notes:
- Internet access to HuggingFace is required to download the optimized model kernels.
- By default, models are stored in:
C:\Users\<USER>\Documents\flm\models\- During installation, you can select a different base folder (e.g., if you choose
C:\Users\<USER>\flm, models will be saved underC:\Users\<USER>\flm\models\).- ⚠️ If HuggingFace is not directly accessible in your region, you can manually download the model (e.g., hf-mirror) and place it in the directory.
🎉🚀 FastFlowLM (FLM) is ready — your NPU is unlocked and you can start chatting with models right away!
Open Task Manager (Ctrl + Shift + Esc). Go to the Performance tab → click NPU to monitor usage.
⚡ Quick Tips:
- Use
/verboseduring a session to turn on performance reporting (toggle off with/verboseagain).- Type
/byeto exit a conversation.- Run
flm listin PowerShell to show all available models.