Giter Club home page Giter Club logo

grok-1's People

Contributors

garethpaul avatar ibab avatar lvelvee avatar mane avatar syzymon avatar xsetech avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

grok-1's Issues

Remove coppersurfer tracker

The tracker tracker.coppersurfer.tk should be removed from the magnet link in the README.

It is no longer a working bittorrent tracker, instead the domain redirects to a parking page.

4bit quantization

Hi,
Thanks for releasing Grok! Is there any chance we could load the model in 4-bit given how large it is? Do you know if bitsandbytes support is planned (cc @TimDettmers)?
Thanks!

Inference memory usage issues

It seems as if the model is being loaded in FP16. I've also noticed how QuantizedWeight8bit is imported in run.py, but not actually used. Is that for runtime quantization with FP16 weights, or is it needed for the released 8bit weights but not actually utilized (due to an oversight?)?

Nr. of devices needed

Running python run.py on a single Nvidia GPU it fails with ValueError: Number of devices 1 must equal the product of mesh_shape (1, 8)

Can the nr of devices be adjusted to 1 only?

ι₯ι₯ι’†ε…ˆ

ι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆι₯ι₯ι’†ε…ˆ

Π Π°Π±ΠΎΡ‚Π°

Илон Маск, возьми Π½Π° Ρ€Π°Π±ΠΎΡ‚Ρƒ поТалуйста )))

Если Π½Π°Π΄ΠΎ Π±ΡƒΠ΄Π΅Ρ‚,Π΄ΠΎΠΎΠ±ΡƒΡ‡Π°ΡŽΡΡŒ. Π“ΠΎΡ‚ΠΎΠ² ΠΊΠΎΠ½Ρ‚Ρ€Π°ΠΊΡ‚ ΠΏΠΎΠ΄ΠΏΠΈΡΠ°Ρ‚ΡŒ Π»Π΅Ρ‚ Π½Π° 20 Ρ…ΠΎΡ‚ΡŒ, Ρ‡Ρ‚ΠΎ послС обучСния ΠΎΡΡ‚Π°Π½ΡƒΡΡŒ)

Π‘Ρ„Π΅Ρ€Π° Π°ΠΉΡ‚ΠΈ, ΠΏΠΈΡˆΡƒ ΠΊΠΎΠ΄ Π½Π° с#. ΠœΠΎΠ³Ρƒ Π΄Ρ€ΡƒΠ³ΠΎΠΉ язык Π²Ρ‹ΡƒΡ‡ΠΈΡ‚ΡŒ. ΠœΠ°Ρ‚Π΅ΠΌ Π΅ΡΡ‚ΡŒ, Ρ„ΠΈΠ·ΠΈΠΊΠ° Π΅ΡΡ‚ΡŒ)

Пиши :)

Original bfloat16 weights

Can we download the original (supposedly bfloat16) weights for fine-tuning? The checkpoint is int8 quantized.

Huggingface Support

Dear xAI team,

Thank you for this gift to the community! Pure awesomeness.

Huggingface support and a Model Card would be appreciated. Huggingface is like a super cool and very popular model hub. This will help the user with a standard workflow.

Thanks,
Vulcan

Run on PC

Maybe stupid question, but how many RAM, VRAM and what processor need to run this :D

Release grok 0 weights

I as a hobbyist am not able to run a 300B model realistically, and fine tuning such a large model is likely even harder. I think Grok 0 weights should be released for people that want to still experiment with Grok but cannot run Grok 1

Egyptian developer

We are living in Egypt we can not have any device that can open this!!

Grok implementation details

not an issue but would be nice if it was in the readme/model.py header:
314B parameters
Mixture of 8 Experts
2 experts used per token
64 layers
48 attention heads for queries
8 attention heads for keys/values
embeddings size: 6,144
rotary embeddings (RoPE)
SentencePiece tokenizer; 131,072 tokens
Supports activation sharding and 8-bit quantization
Max seq length (context): 8,192 tokens

Opensource training pipeline

Dear xAI team

Would xAI consider opensourcing the training pipleline or recipe to reproduce models? As I understand the weights and inference part is open. Arguably this is not fully opensource.

The benefits of open sourcing the training pipeline is mutual, researches could experiment and contribute to a more effective training pipeline or model improvements... xAI could benefit from that to get more market share and be the true Open AI!

πŸ€– Now run grok-1 with less than πŸ”² 420 G VRAM ⚑

Surprise! You can't run it on your average desktop or laptop!

Run grok-1 with less than πŸ”² 420 G VRAM

Run grok on an Mac Studio with an M2 Ultra and 192GB of unified ram: See: llama2.cpp / grok-1 support
@ibab_ml on X

You need a beefy machine to run grok-1

Grok-1 is a true mystical creature. Rumor has it that it lives in the cores of 8 GPU's and that the Model must fit in the VRAM.

This implies that you need a very beefy machine. Very very beefy machine. So beefy...

How do you know if your machine is beefy or not?

Your machine is not beefy if it is not big - the bigger the better, size matters! It has to make the sound of a jet engine when it thinks, also it has to be hot to the touch mostly.

It must also smell like burnt plastic at times. The more big iron, the more heavy the more beefy! If you didn't pay a heavy price for it, such as 100k$++, an arm and a leg, then it is not beefy.

What are some of the working setups?

llama2.cpp:
Mac
  • Mac Studio with an M2 Ultra
  • 192GB of unified ram.
AMD
  • Threadripper 3955WX
  • 256GB RAM
  • 0.5 tokens per second.
This repo:
Intel + Nvidia
  • GPU: 8 x A100 80G
  • Total VRAM: 640G
  • CPU: 2 x Xeon 8480+
  • RAM: 1.5 TB

#168 (comment)

AMD
  • GPU: 8 x Instinct MI300X GPU 190G
  • Total VRAM: 1520G

#130 (comment)

Other / Container / Cloud
  • GPU: 8 x A100 80G
  • Total VRAM: 640G
  • K8 cluster

#6 (comment)

What can you do about it?

Try: See: llama2.cpp / grok-1 support

What are the other options?

  • Rent a GPU cloud instance with sufficient resources
  • Subscribe to grok at X (twitter.com)
  • Study the blade, save up money
  • Get someone to cosplay as grok

What is the Answer to the Ultimate Question of Life, the Universe, and Everything?

#42

Ref:
#168 (comment)
#130 (comment)
#130 (comment)
#125 (comment)
#6 (comment)
ggerganov/llama.cpp#6204 (comment)
ggerganov/llama.cpp#6204 (comment)
ggerganov/llama.cpp#6204 (comment)

See: Discussion
Note: This issue has been edited totally to elevate Issue 42 to serve a much better cause. @xSetech Would you not be tempted to pin this?
Edit: Corrected llama2.cpp inaccuracies

Error

C:\Users\MARKELOX\Desktop\grok-main>py run.py
INFO:jax._src.xla_bridge:Unable to initialize backend 'cuda':
INFO:jax._src.xla_bridge:Unable to initialize backend 'rocm': module 'jaxlib.xla_extension' has no attribute 'GpuAllocatorConfig'
INFO:jax._src.xla_bridge:Unable to initialize backend 'tpu': UNIMPLEMENTED: LoadPjrtPlugin is not implemented on windows yet.
INFO:rank:Initializing mesh for self.local_mesh_config=(1, 8) self.between_hosts_config=(1, 1)...
INFO:rank:Detected 1 devices in mesh
Traceback (most recent call last):
File "C:\Users\MARKELOX\Desktop\grok-main\run.py", line 72, in
main()
File "C:\Users\MARKELOX\Desktop\grok-main\run.py", line 63, in main
inference_runner.initialize()
File "C:\Users\MARKELOX\Desktop\grok-main\runners.py", line 282, in initialize
runner.initialize(
File "C:\Users\MARKELOX\Desktop\grok-main\runners.py", line 181, in initialize
self.mesh = make_mesh(self.local_mesh_config, self.between_hosts_config)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\MARKELOX\Desktop\grok-main\runners.py", line 586, in make_mesh
device_mesh = mesh_utils.create_hybrid_device_mesh(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\MARKELOX\AppData\Local\Programs\Python\Python312\Lib\site-packages\jax\experimental\mesh_utils.py", line 373, in create_hybrid_device_mesh
per_granule_meshes = [create_device_mesh(mesh_shape, granule)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\MARKELOX\AppData\Local\Programs\Python\Python312\Lib\site-packages\jax\experimental\mesh_utils.py", line 302, in create_device_mesh
raise ValueError(f'Number of devices {len(devices)} must equal the product '
ValueError: Number of devices 1 must equal the product of mesh_shape (1, 8)

Can i run grok in gcolab?

I tried to load this model in Google colab but it fails to load, is there any way that it will get optimized or something to load in Google colab or any other notebook like paperspace?

Error "Could not find a version that satisfies the requirement jaxlib==0.4.25+cuda12.cudnn89"

Hello everyone.

I've tried to run the pip install, but I'm facing the following error:

ERROR: Could not find a version that satisfies the requirement jaxlib==0.4.25+cuda12.cudnn89; extra == "cuda12_pip" (from jax[cuda12-pip]) (from versions: 0.4.17, 0.4.18, 0.4.19, 0.4.20, 0.4.21, 0.4.22, 0.4.23, 0.4.24, 0.4.25)
ERROR: No matching distribution found for jaxlib==0.4.25+cuda12.cudnn89; extra == "cuda12_pip"

I'm install on MacOsX and in a Ubuntu and faced the same issue.

Anyone else got the same error?

Error when installing requirements

i have installed python 3.10 and venv. Trying to "pip install -r requirements.txt"

ERROR: Ignored the following versions that require a different python version: 1.6.2 Requires-Python >=3.7,<3.10; 1.6.3 Requires-Python >=3.7,<3.10; 1.7.0 Requires-Python >=3.7,<3.10; 1.7.1 Requires-Python >=3.7,<3.10
ERROR: Could not find a version that satisfies the requirement jaxlib==0.4.25+cuda12.cudnn89; extra == "cuda12_pip" (from jax[cuda12-pip]) (from versions: 0.4.13, 0.4.14, 0.4.16, 0.4.17, 0.4.18, 0.4.19, 0.4.20, 0.4.21, 0.4.22, 0.4.23, 0.4.24, 0.4.25)
ERROR: No matching distribution found for jaxlib==0.4.25+cuda12.cudnn89; extra == "cuda12_pip"

Is there a scientific paper?

Is there a scientific paper accompanying this release? I've searched but couldn't find one. I find it odd that the weights would be released but not the research.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    πŸ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. πŸ“ŠπŸ“ˆπŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❀️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.