iarroyof commited on
Commit
4732ef5
·
verified ·
1 Parent(s): 8eef088

Update README.md

Browse files

Update usage sample code

Files changed (1) hide show
  1. README.md +5 -6
README.md CHANGED
@@ -36,18 +36,17 @@ This model can be used for text-to-text generation tasks like question answering
36
 
37
  ```python
38
  from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
 
39
 
40
- tokenizer = AutoTokenizer.from_pretrained("iarroyof/t5-11b-ssm-nq-sharded")
41
  model = AutoModelForSeq2SeqLM.from_pretrained(
42
- "iarroyof/t5-11b-ssm-nq-sharded",
43
- device_map="auto",
44
- max_memory={0: "40GB", 1: "40GB", "cpu": "30GB"},
45
  low_cpu_mem_usage=True,
46
  torch_dtype=torch.float16,
47
- trust_remote_code=True
48
  )
49
 
50
- inputs = tokenizer("Translate English to French: How are you?", return_tensors="pt").input_ids
51
  outputs = model.generate(inputs)
52
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
53
  ---
 
36
 
37
  ```python
38
  from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
39
+ import torch
40
 
41
+ tokenizer = AutoTokenizer.from_pretrained('iarroyof/t5-11b-ssm-nq-sharded')
42
  model = AutoModelForSeq2SeqLM.from_pretrained(
43
+ 'iarroyof/t5-11b-ssm-nq-sharded',
44
+ device_map='auto',
 
45
  low_cpu_mem_usage=True,
46
  torch_dtype=torch.float16,
 
47
  )
48
 
49
+ inputs = tokenizer('What is and how to deal with insomnia?', return_tensors='pt').input_ids.to('cuda')
50
  outputs = model.generate(inputs)
51
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
52
  ---