Comments (3)
lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py Lines 608-609, swap the annotation.
from vila.
lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py Lines 608-609, swap the annotation.
Thanks, it worked for my GPU now! However, the output is really weird; it outputs a meaningless string of empty spaces and commas. I faced the same issue with another Vision Language Model, while some other Vision Language Models work well. I believe this might be due to the transformers library version. Anyway, I also tried running VILA on the CPU, and in that case, it worked fine.
from vila.
lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py Lines 608-609, swap the annotation.
Thanks, it worked for my GPU now! However, the output is really weird; it outputs a meaningless string of empty spaces and commas. I faced the same issue with another Vision Language Model, while some other Vision Language Models work well. I believe this might be due to the transformers library version. Anyway, I also tried running VILA on the CPU, and in that case, it worked fine.
Me, too. I've had similar issues with redundant commas and spaces. However, when I use the VILA1.5-3B model to input a video along with some questions, it actually performs better than the 8B model. Sometimes it generates coherent responses, but other times it only replies with one to three words.
from vila.
Related Issues (20)
- How to convert model to gguf HOT 3
- Deployment to SageMaker and/or HuggingFace Inference Endpoints Fails With Error HOT 5
- Whether the visual encoder participates in training HOT 1
- Support for multi-video captioning with multiple grid image inputs? HOT 1
- Multi-Image or Multi-Video Inference Example HOT 2
- question: what does 'repack_multimodal_data' function do? HOT 1
- release schedule for the "VILA1.5-34b-4bit-AWQ" model. HOT 1
- Is there any way to increase the context window? HOT 4
- Question re. LanguageModel vs LanguageModelForCausalLM functionalies HOT 2
- Llama2 or Llama3
- What is the conv_mode for VILA-1.5-3B ? HOT 1
- AttributeError: 'Image' object has no attribute 'shape' HOT 6
- Support VILA with lmdeploy
- Image text retrieval support
- COYO-700M Dataset Download Script Error
- [Help] Using VILA1.5-40b model for Video Descriptions
- [HELP] Do we have any docker image for Jetson platform ?
- No training scripts in scripts/v1_5/paper/
- About sharegpt_video. How do you make video file from jpeg images?
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from vila.