Open
Description
Description of the bug:
I exported my Llama3.2-BB to .task
using:
# pip installs
pip install --upgrade pip setuptools wheel
pip install ai-edge-torch-nightly
pip install ai-edge-quantizer-nightly
pip install transformers
pip install tensorflow-cpu
# tflite & task export
curl -o verify.py https://raw.githubusercontent.com/google-ai-edge/ai-edge-torch/refs/heads/main/ai_edge_torch/generative/examples/llama/verify.py
curl -o convert_to_tflite.py https://raw.githubusercontent.com/google-ai-edge/ai-edge-torch/refs/heads/main/ai_edge_torch/generative/examples/llama/convert_to_tflite.py
curl -o tokenizer_to_sentencepiece.py https://raw.githubusercontent.com/google-ai-edge/ai-edge-torch/main/ai_edge_torch/generative/tools/tokenizer_to_sentencepiece.py
python3 verify.py
python3 tokenizer_to_sentencepiece.py --checkpoint=meta-llama/Llama-3.2-1B-Instruct --output_path=/tmp/llama3.spm.model
python3 convert_to_tflite.py --checkpoint_path '/root/.cache/huggingface/hub/models--meta-llama--Llama-3.2-1B-Instruct/snapshots/9213176726f574b556790deb65791e0c5aa438b6'
python3 -c "from mediapipe.tasks.python.genai import bundler; \
config = bundler.BundleConfig(tflite_model='/tmp/llama_1b_q8_ekv1280.tflite', \
tokenizer_model='/tmp/llama3.spm.model', \
start_token='<|begin_of_text|>', \
stop_tokens=['<|end_of_text|>'], \
output_filename='/tmp/llama_1b_q8_ekv1280.task', \
enable_bytes_to_unicode_mapping=False); \
print('Configuration created:', config); \
bundler.create_bundle(config)"
When loading the .task
model into into the mediapipe
app I get:
Error -
internal: Failed to initialize session: %sINTERNAL: CalculatorGraph::Run() failed: Calculator::Open() for node "odml.infra.TfLitePrefillDecodeRunnerCalculator" failed; RET_CHECK failure (external/odml/odml/infra/genai/inference/utils/tflite_utils/tflite_llm_utils.cc:59) std::find_if(signature_keys.begin(), signature_keys.end(), [&](const std::string* key) { return *key == required_key; }) != signature_keys.end()
Am I missing something?
Actual vs expected behavior:
No response
Any other information you'd like to share?
No response