Files
MultimodalOCR/models/OpenFlamingo/__pycache__/OpenFlamingo.cpython-310.pyc

17 lines
2.0 KiB
Plaintext
Raw Normal View History

2023-05-27 17:21:39 +08:00
o
g<>ld<6C><00>@s^ddlmZddlmZddlZddlmZmZddlm Z ddl
Z
dd<08>Z Gd d
<EFBFBD>d
<EFBFBD>Z dS) <0B>)<01>create_model_and_transforms)<01>hf_hub_downloadN<64>)<02> pad_image<67> resize_image)<01>ImagecCst<00>d|d<02>dS)NzQuestion|Answer<65>r)<02>re<72>split)<01> predictions<6E>r <00>C/home/zhangli/GPT4/MutimodelOCR/models/OpenFlamingo/OpenFlamingo.py<70>postprocess_vqa_generationsrc@s eZdZddd<04>Zd dd<07>ZdS)
<EFBFBD> OpenFlamingo<67>returnNcCsVtdd||dd<04>\}}}tj|dd<06>}|j|dd<08>|<04>|<03>|_||_||_||_dS) NzViT-L-14<31>openai<61>)Zclip_vision_encoder_pathZclip_vision_encoder_pretrainedZlang_encoder_pathZtokenizer_pathZcross_attn_every_n_layers<72>cpu)<01> map_locationF)<01>strict) r<00>torch<63>load<61>load_state_dict<63>to<74>model<65>image_processor<6F> tokenizer<65>device)<08>self<6C>
llama_path<EFBFBD> check_pointrrrr<00>
checkpointr r r <00>__init__
s <0C> 
zOpenFlamingo.__init__<5F>resizec Cs<>d|j_|jd|<02>d<03>gdd<05><02>|j<03>}t|dd<00>}t<05>|<01>}|dkr*t|d <09>}|<00>|<01><01> d<07>g}t
j |dd
<EFBFBD>}|<06> d <0B><01> d<07><01>|j<03>}|j j ||d|d d dd<0F>}|jj|d|d<00>dd<11>} |S)N<>leftz<image>Question:z Answer:<3A>pt)<01>return_tensors<72> input_idsrr#)<02><>r()<01>dimr<00>attention_mask<73>0<00>)<05>vision_x<5F>lang_xr*<00>max_new_tokens<6E> num_beamsT)<01>skip_special_tokens)r<00> padding_siderr<00>lenr<00>openrr<00> unsqueezer<00>catr<00>generate<74>decode) r<00>image<67>question<6F>namer.Z len_inputr-Zgenerated_text<78>answerr r r r7s0 <04><02>

<06>zOpenFlamingo.generate)rN)r#)<05>__name__<5F>
__module__<EFBFBD> __qualname__r"r7r r r r r s
r) Z open_flamingor<00>huggingface_hubrr<00>processrr<00>PILrr rrr r r r <00><module>s