YFanwang commited on
Commit
ab64070
·
verified ·
1 Parent(s): 490436d

Upload folder using huggingface_hub

Browse files
Files changed (45) hide show
  1. logs_oct11/LICENSE +201 -0
  2. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation_20251012_155204.log +0 -0
  3. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.1_2e-1_connector-1.0_2.1_2e-1_ablation_20251012_162128.log +0 -0
  4. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation_20251012_164611.log +219 -0
  5. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.5_2e-1_connector-1.0_2.5_2e-1_ablation_20251012_164737.log +0 -0
  6. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.7_2e-1_connector-1.0_2.7_2e-1_ablation_20251012_171238.log +0 -0
  7. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation_20251012_173957.log +219 -0
  8. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_1e-2_connector-3.0_0.5_1e-2_ablation_20251012_174122.log +0 -0
  9. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation_20251012_180629.log +219 -0
  10. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_3e-2_connector-3.0_0.5_3e-2_ablation_20251012_180754.log +0 -0
  11. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-1_connector-3.0_0.5_5e-1_ablation_20251012_183317.log +0 -0
  12. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation_20251012_190124.log +219 -0
  13. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-1_connector-3.0_0.5_7e-1_ablation_20251012_190249.log +0 -0
  14. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-2_connector-3.0_0.5_7e-2_ablation_20251012_192708.log +0 -0
  15. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-1_connector-3.0_0.5_9e-1_ablation_20251012_195147.log +0 -0
  16. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-2_connector-3.0_0.5_9e-2_ablation_20251012_202158.log +0 -0
  17. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.7_2e-1_connector-3.0_0.7_2e-1_ablation_20251012_204649.log +0 -0
  18. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.9_2e-1_connector-3.0_0.9_2e-1_ablation_20251012_211112.log +0 -0
  19. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.1_2e-1_connector-3.0_1.1_2e-1_ablation_20251012_213606.log +0 -0
  20. logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation_20251012_220114.log +219 -0
  21. logs_oct11/logs_oct10.tar +3 -0
  22. logs_oct11/logs_oct9.tar +3 -0
  23. logs_oct11/plot_ablation.py +120 -0
  24. logs_oct11/pretrain_qwen.zip +3 -0
  25. logs_oct11/pyproject.toml +38 -0
  26. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation_20251012_052820.log +0 -0
  27. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation_20251012_155158.log +6 -0
  28. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.1_2e-1_connector-1.0_2.1_2e-1_ablation_20251012_060227.log +0 -0
  29. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation_20251012_081931.log +0 -0
  30. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.5_2e-1_connector-1.0_2.5_2e-1_ablation_20251012_090245.log +0 -0
  31. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.7_2e-1_connector-1.0_2.7_2e-1_ablation_20251012_100457.log +0 -0
  32. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation_20251012_113628.log +0 -0
  33. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_1e-2_connector-3.0_0.5_1e-2_ablation_20251012_091538.log +0 -0
  34. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation_20251012_123546.log +0 -0
  35. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_3e-2_connector-3.0_0.5_3e-2_ablation_20251012_100541.log +0 -0
  36. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-1_connector-3.0_0.5_5e-1_ablation_20251012_052751.log +0 -0
  37. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation_20251012_113222.log +0 -0
  38. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-1_connector-3.0_0.5_7e-1_ablation_20251012_071236.log +0 -0
  39. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-2_connector-3.0_0.5_7e-2_ablation_20251012_132440.log +0 -0
  40. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-1_connector-3.0_0.5_9e-1_ablation_20251012_081416.log +0 -0
  41. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-2_connector-3.0_0.5_9e-2_ablation_20251012_142902.log +0 -0
  42. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.7_2e-1_connector-3.0_0.7_2e-1_ablation_20251012_133702.log +0 -0
  43. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.9_2e-1_connector-3.0_0.9_2e-1_ablation_20251012_144110.log +0 -0
  44. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.1_2e-1_connector-3.0_1.1_2e-1_ablation_20251012_151526.log +0 -0
  45. logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation_20251012_154943.log +169 -0
logs_oct11/LICENSE ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Apache License
2
+ Version 2.0, January 2004
3
+ http://www.apache.org/licenses/
4
+
5
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
6
+
7
+ 1. Definitions.
8
+
9
+ "License" shall mean the terms and conditions for use, reproduction,
10
+ and distribution as defined by Sections 1 through 9 of this document.
11
+
12
+ "Licensor" shall mean the copyright owner or entity authorized by
13
+ the copyright owner that is granting the License.
14
+
15
+ "Legal Entity" shall mean the union of the acting entity and all
16
+ other entities that control, are controlled by, or are under common
17
+ control with that entity. For the purposes of this definition,
18
+ "control" means (i) the power, direct or indirect, to cause the
19
+ direction or management of such entity, whether by contract or
20
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
21
+ outstanding shares, or (iii) beneficial ownership of such entity.
22
+
23
+ "You" (or "Your") shall mean an individual or Legal Entity
24
+ exercising permissions granted by this License.
25
+
26
+ "Source" form shall mean the preferred form for making modifications,
27
+ including but not limited to software source code, documentation
28
+ source, and configuration files.
29
+
30
+ "Object" form shall mean any form resulting from mechanical
31
+ transformation or translation of a Source form, including but
32
+ not limited to compiled object code, generated documentation,
33
+ and conversions to other media types.
34
+
35
+ "Work" shall mean the work of authorship, whether in Source or
36
+ Object form, made available under the License, as indicated by a
37
+ copyright notice that is included in or attached to the work
38
+ (an example is provided in the Appendix below).
39
+
40
+ "Derivative Works" shall mean any work, whether in Source or Object
41
+ form, that is based on (or derived from) the Work and for which the
42
+ editorial revisions, annotations, elaborations, or other modifications
43
+ represent, as a whole, an original work of authorship. For the purposes
44
+ of this License, Derivative Works shall not include works that remain
45
+ separable from, or merely link (or bind by name) to the interfaces of,
46
+ the Work and Derivative Works thereof.
47
+
48
+ "Contribution" shall mean any work of authorship, including
49
+ the original version of the Work and any modifications or additions
50
+ to that Work or Derivative Works thereof, that is intentionally
51
+ submitted to Licensor for inclusion in the Work by the copyright owner
52
+ or by an individual or Legal Entity authorized to submit on behalf of
53
+ the copyright owner. For the purposes of this definition, "submitted"
54
+ means any form of electronic, verbal, or written communication sent
55
+ to the Licensor or its representatives, including but not limited to
56
+ communication on electronic mailing lists, source code control systems,
57
+ and issue tracking systems that are managed by, or on behalf of, the
58
+ Licensor for the purpose of discussing and improving the Work, but
59
+ excluding communication that is conspicuously marked or otherwise
60
+ designated in writing by the copyright owner as "Not a Contribution."
61
+
62
+ "Contributor" shall mean Licensor and any individual or Legal Entity
63
+ on behalf of whom a Contribution has been received by Licensor and
64
+ subsequently incorporated within the Work.
65
+
66
+ 2. Grant of Copyright License. Subject to the terms and conditions of
67
+ this License, each Contributor hereby grants to You a perpetual,
68
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
69
+ copyright license to reproduce, prepare Derivative Works of,
70
+ publicly display, publicly perform, sublicense, and distribute the
71
+ Work and such Derivative Works in Source or Object form.
72
+
73
+ 3. Grant of Patent License. Subject to the terms and conditions of
74
+ this License, each Contributor hereby grants to You a perpetual,
75
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
76
+ (except as stated in this section) patent license to make, have made,
77
+ use, offer to sell, sell, import, and otherwise transfer the Work,
78
+ where such license applies only to those patent claims licensable
79
+ by such Contributor that are necessarily infringed by their
80
+ Contribution(s) alone or by combination of their Contribution(s)
81
+ with the Work to which such Contribution(s) was submitted. If You
82
+ institute patent litigation against any entity (including a
83
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
84
+ or a Contribution incorporated within the Work constitutes direct
85
+ or contributory patent infringement, then any patent licenses
86
+ granted to You under this License for that Work shall terminate
87
+ as of the date such litigation is filed.
88
+
89
+ 4. Redistribution. You may reproduce and distribute copies of the
90
+ Work or Derivative Works thereof in any medium, with or without
91
+ modifications, and in Source or Object form, provided that You
92
+ meet the following conditions:
93
+
94
+ (a) You must give any other recipients of the Work or
95
+ Derivative Works a copy of this License; and
96
+
97
+ (b) You must cause any modified files to carry prominent notices
98
+ stating that You changed the files; and
99
+
100
+ (c) You must retain, in the Source form of any Derivative Works
101
+ that You distribute, all copyright, patent, trademark, and
102
+ attribution notices from the Source form of the Work,
103
+ excluding those notices that do not pertain to any part of
104
+ the Derivative Works; and
105
+
106
+ (d) If the Work includes a "NOTICE" text file as part of its
107
+ distribution, then any Derivative Works that You distribute must
108
+ include a readable copy of the attribution notices contained
109
+ within such NOTICE file, excluding those notices that do not
110
+ pertain to any part of the Derivative Works, in at least one
111
+ of the following places: within a NOTICE text file distributed
112
+ as part of the Derivative Works; within the Source form or
113
+ documentation, if provided along with the Derivative Works; or,
114
+ within a display generated by the Derivative Works, if and
115
+ wherever such third-party notices normally appear. The contents
116
+ of the NOTICE file are for informational purposes only and
117
+ do not modify the License. You may add Your own attribution
118
+ notices within Derivative Works that You distribute, alongside
119
+ or as an addendum to the NOTICE text from the Work, provided
120
+ that such additional attribution notices cannot be construed
121
+ as modifying the License.
122
+
123
+ You may add Your own copyright statement to Your modifications and
124
+ may provide additional or different license terms and conditions
125
+ for use, reproduction, or distribution of Your modifications, or
126
+ for any such Derivative Works as a whole, provided Your use,
127
+ reproduction, and distribution of the Work otherwise complies with
128
+ the conditions stated in this License.
129
+
130
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
131
+ any Contribution intentionally submitted for inclusion in the Work
132
+ by You to the Licensor shall be under the terms and conditions of
133
+ this License, without any additional terms or conditions.
134
+ Notwithstanding the above, nothing herein shall supersede or modify
135
+ the terms of any separate license agreement you may have executed
136
+ with Licensor regarding such Contributions.
137
+
138
+ 6. Trademarks. This License does not grant permission to use the trade
139
+ names, trademarks, service marks, or product names of the Licensor,
140
+ except as required for reasonable and customary use in describing the
141
+ origin of the Work and reproducing the content of the NOTICE file.
142
+
143
+ 7. Disclaimer of Warranty. Unless required by applicable law or
144
+ agreed to in writing, Licensor provides the Work (and each
145
+ Contributor provides its Contributions) on an "AS IS" BASIS,
146
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
147
+ implied, including, without limitation, any warranties or conditions
148
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
149
+ PARTICULAR PURPOSE. You are solely responsible for determining the
150
+ appropriateness of using or redistributing the Work and assume any
151
+ risks associated with Your exercise of permissions under this License.
152
+
153
+ 8. Limitation of Liability. In no event and under no legal theory,
154
+ whether in tort (including negligence), contract, or otherwise,
155
+ unless required by applicable law (such as deliberate and grossly
156
+ negligent acts) or agreed to in writing, shall any Contributor be
157
+ liable to You for damages, including any direct, indirect, special,
158
+ incidental, or consequential damages of any character arising as a
159
+ result of this License or out of the use or inability to use the
160
+ Work (including but not limited to damages for loss of goodwill,
161
+ work stoppage, computer failure or malfunction, or any and all
162
+ other commercial damages or losses), even if such Contributor
163
+ has been advised of the possibility of such damages.
164
+
165
+ 9. Accepting Warranty or Additional Liability. While redistributing
166
+ the Work or Derivative Works thereof, You may choose to offer,
167
+ and charge a fee for, acceptance of support, warranty, indemnity,
168
+ or other liability obligations and/or rights consistent with this
169
+ License. However, in accepting such obligations, You may act only
170
+ on Your own behalf and on Your sole responsibility, not on behalf
171
+ of any other Contributor, and only if You agree to indemnify,
172
+ defend, and hold each Contributor harmless for any liability
173
+ incurred by, or claims asserted against, such Contributor by reason
174
+ of your accepting any such warranty or additional liability.
175
+
176
+ END OF TERMS AND CONDITIONS
177
+
178
+ APPENDIX: How to apply the Apache License to your work.
179
+
180
+ To apply the Apache License to your work, attach the following
181
+ boilerplate notice, with the fields enclosed by brackets "[]"
182
+ replaced with your own identifying information. (Don't include
183
+ the brackets!) The text should be enclosed in the appropriate
184
+ comment syntax for the file format. We also recommend that a
185
+ file or class name and description of purpose be included on the
186
+ same "printed page" as the copyright notice for easier
187
+ identification within third-party archives.
188
+
189
+ Copyright [TinyLLaVA]
190
+
191
+ Licensed under the Apache License, Version 2.0 (the "License");
192
+ you may not use this file except in compliance with the License.
193
+ You may obtain a copy of the License at
194
+
195
+ http://www.apache.org/licenses/LICENSE-2.0
196
+
197
+ Unless required by applicable law or agreed to in writing, software
198
+ distributed under the License is distributed on an "AS IS" BASIS,
199
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
200
+ See the License for the specific language governing permissions and
201
+ limitations under the License.
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation_20251012_155204.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.1_2e-1_connector-1.0_2.1_2e-1_ablation_20251012_162128.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation_20251012_164611.log ADDED
@@ -0,0 +1,219 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==== STARTING EXPERIMENT: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation ====
2
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation_20251012_164611.log
3
+ Timestamp: 2025-10-12 16:46:11
4
+ =====================================
5
+ Processing: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation
6
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
7
+ import pynvml # type: ignore[import]
8
+ [2025-10-12 16:46:14,468] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ Traceback (most recent call last):
10
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 488, in <module>
11
+ main()
12
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 123, in main
13
+ config_mask = TinyLlavaConfig.from_pretrained(model_args.mask_model_name_or_path)
14
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
15
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
16
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
17
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
18
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
19
+ resolved_config_file = cached_file(
20
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 369, in cached_file
21
+ raise EnvironmentError(
22
+ OSError: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation does not appear to have a file named config.json. Checkout 'https://huggingface.co//nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/tree/main' for available files.
23
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
24
+ import pynvml # type: ignore[import]
25
+ [2025-10-12 16:46:21,768] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
26
+ Traceback (most recent call last):
27
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
28
+ resolved_file = hf_hub_download(
29
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
30
+ validate_repo_id(arg_value)
31
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
32
+ raise HFValidationError(
33
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
34
+
35
+ The above exception was the direct cause of the following exception:
36
+
37
+ Traceback (most recent call last):
38
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
39
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
40
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
41
+ resolved_config_file = cached_file(
42
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
43
+ raise EnvironmentError(
44
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
45
+
46
+ During handling of the above exception, another exception occurred:
47
+
48
+ Traceback (most recent call last):
49
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
50
+ resolved_file = hf_hub_download(
51
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
52
+ validate_repo_id(arg_value)
53
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
54
+ raise HFValidationError(
55
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
56
+
57
+ The above exception was the direct cause of the following exception:
58
+
59
+ Traceback (most recent call last):
60
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
61
+ return _run_code(code, main_globals, None,
62
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
63
+ exec(code, run_globals)
64
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 180, in <module>
65
+ eval_model(args)
66
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 88, in eval_model
67
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
68
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
69
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
70
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
71
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
72
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
73
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
74
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
75
+ resolved_config_file = cached_file(
76
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
77
+ raise EnvironmentError(
78
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
79
+ Traceback (most recent call last):
80
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 31, in <module>
81
+ eval_model(args)
82
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 7, in eval_model
83
+ answers = [json.loads(q) for q in open(os.path.expanduser(args.answers_file), "r")]
84
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation-mask_applied.jsonl'
85
+ Traceback (most recent call last):
86
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MMMU/eval/main_eval_only.py", line 19, in <module>
87
+ output_dict = json.load(open(args.output_path))
88
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation-mask_applied_output.json'
89
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
90
+ import pynvml # type: ignore[import]
91
+ [2025-10-12 16:46:28,066] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
92
+ Traceback (most recent call last):
93
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
94
+ resolved_file = hf_hub_download(
95
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
96
+ validate_repo_id(arg_value)
97
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
98
+ raise HFValidationError(
99
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
100
+
101
+ The above exception was the direct cause of the following exception:
102
+
103
+ Traceback (most recent call last):
104
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
105
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
106
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
107
+ resolved_config_file = cached_file(
108
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
109
+ raise EnvironmentError(
110
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
111
+
112
+ During handling of the above exception, another exception occurred:
113
+
114
+ Traceback (most recent call last):
115
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
116
+ resolved_file = hf_hub_download(
117
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
118
+ validate_repo_id(arg_value)
119
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
120
+ raise HFValidationError(
121
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
122
+
123
+ The above exception was the direct cause of the following exception:
124
+
125
+ Traceback (most recent call last):
126
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
127
+ return _run_code(code, main_globals, None,
128
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
129
+ exec(code, run_globals)
130
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 146, in <module>
131
+ eval_model(args)
132
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 79, in eval_model
133
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
134
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
135
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
136
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
137
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
138
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
139
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
140
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
141
+ resolved_config_file = cached_file(
142
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
143
+ raise EnvironmentError(
144
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
145
+ Traceback (most recent call last):
146
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/convert_answer_to_mme.py", line 52, in <module>
147
+ answers = [json.loads(line) for line in open(os.path.join('answers', f'{experiment}.jsonl'))]
148
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation-mask_applied.jsonl'
149
+ =========== Perception ===========
150
+ Traceback (most recent call last):
151
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 164, in <module>
152
+ cal.process_result(results_dir)
153
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 98, in process_result
154
+ lines = open(task_txt, 'r').readlines()
155
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation-mask_applied/existence.txt'
156
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
157
+ import pynvml # type: ignore[import]
158
+ [2025-10-12 16:47:34,409] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
159
+ Traceback (most recent call last):
160
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
161
+ resolved_file = hf_hub_download(
162
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
163
+ validate_repo_id(arg_value)
164
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
165
+ raise HFValidationError(
166
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
167
+
168
+ The above exception was the direct cause of the following exception:
169
+
170
+ Traceback (most recent call last):
171
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
172
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
173
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
174
+ resolved_config_file = cached_file(
175
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
176
+ raise EnvironmentError(
177
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
178
+
179
+ During handling of the above exception, another exception occurred:
180
+
181
+ Traceback (most recent call last):
182
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
183
+ resolved_file = hf_hub_download(
184
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
185
+ validate_repo_id(arg_value)
186
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
187
+ raise HFValidationError(
188
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
189
+
190
+ The above exception was the direct cause of the following exception:
191
+
192
+ Traceback (most recent call last):
193
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
194
+ return _run_code(code, main_globals, None,
195
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
196
+ exec(code, run_globals)
197
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 108, in <module>
198
+ eval_model(args)
199
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 31, in eval_model
200
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
201
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
202
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
203
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
204
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
205
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
206
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
207
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
208
+ resolved_config_file = cached_file(
209
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
210
+ raise EnvironmentError(
211
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
212
+ Traceback (most recent call last):
213
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/eval_science_qa.py", line 45, in <module>
214
+ predictions = [json.loads(line) for line in open(args.result_file)]
215
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/scienceqa/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation-mask_applied.jsonl'
216
+ ==== EXPERIMENT COMPLETED: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation ====
217
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation_20251012_164611.log
218
+ Timestamp: 2025-10-12 16:47:37
219
+ =====================================
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.5_2e-1_connector-1.0_2.5_2e-1_ablation_20251012_164737.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.7_2e-1_connector-1.0_2.7_2e-1_ablation_20251012_171238.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation_20251012_173957.log ADDED
@@ -0,0 +1,219 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==== STARTING EXPERIMENT: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation ====
2
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation_20251012_173957.log
3
+ Timestamp: 2025-10-12 17:39:57
4
+ =====================================
5
+ Processing: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation
6
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
7
+ import pynvml # type: ignore[import]
8
+ [2025-10-12 17:40:00,477] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ Traceback (most recent call last):
10
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 488, in <module>
11
+ main()
12
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 123, in main
13
+ config_mask = TinyLlavaConfig.from_pretrained(model_args.mask_model_name_or_path)
14
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
15
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
16
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
17
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
18
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
19
+ resolved_config_file = cached_file(
20
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 369, in cached_file
21
+ raise EnvironmentError(
22
+ OSError: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation does not appear to have a file named config.json. Checkout 'https://huggingface.co//nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/tree/main' for available files.
23
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
24
+ import pynvml # type: ignore[import]
25
+ [2025-10-12 17:40:08,208] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
26
+ Traceback (most recent call last):
27
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
28
+ resolved_file = hf_hub_download(
29
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
30
+ validate_repo_id(arg_value)
31
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
32
+ raise HFValidationError(
33
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
34
+
35
+ The above exception was the direct cause of the following exception:
36
+
37
+ Traceback (most recent call last):
38
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
39
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
40
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
41
+ resolved_config_file = cached_file(
42
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
43
+ raise EnvironmentError(
44
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
45
+
46
+ During handling of the above exception, another exception occurred:
47
+
48
+ Traceback (most recent call last):
49
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
50
+ resolved_file = hf_hub_download(
51
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
52
+ validate_repo_id(arg_value)
53
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
54
+ raise HFValidationError(
55
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
56
+
57
+ The above exception was the direct cause of the following exception:
58
+
59
+ Traceback (most recent call last):
60
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
61
+ return _run_code(code, main_globals, None,
62
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
63
+ exec(code, run_globals)
64
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 180, in <module>
65
+ eval_model(args)
66
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 88, in eval_model
67
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
68
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
69
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
70
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
71
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
72
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
73
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
74
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
75
+ resolved_config_file = cached_file(
76
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
77
+ raise EnvironmentError(
78
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
79
+ Traceback (most recent call last):
80
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 31, in <module>
81
+ eval_model(args)
82
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 7, in eval_model
83
+ answers = [json.loads(q) for q in open(os.path.expanduser(args.answers_file), "r")]
84
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation-mask_applied.jsonl'
85
+ Traceback (most recent call last):
86
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MMMU/eval/main_eval_only.py", line 19, in <module>
87
+ output_dict = json.load(open(args.output_path))
88
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation-mask_applied_output.json'
89
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
90
+ import pynvml # type: ignore[import]
91
+ [2025-10-12 17:40:14,451] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
92
+ Traceback (most recent call last):
93
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
94
+ resolved_file = hf_hub_download(
95
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
96
+ validate_repo_id(arg_value)
97
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
98
+ raise HFValidationError(
99
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
100
+
101
+ The above exception was the direct cause of the following exception:
102
+
103
+ Traceback (most recent call last):
104
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
105
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
106
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
107
+ resolved_config_file = cached_file(
108
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
109
+ raise EnvironmentError(
110
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
111
+
112
+ During handling of the above exception, another exception occurred:
113
+
114
+ Traceback (most recent call last):
115
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
116
+ resolved_file = hf_hub_download(
117
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
118
+ validate_repo_id(arg_value)
119
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
120
+ raise HFValidationError(
121
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
122
+
123
+ The above exception was the direct cause of the following exception:
124
+
125
+ Traceback (most recent call last):
126
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
127
+ return _run_code(code, main_globals, None,
128
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
129
+ exec(code, run_globals)
130
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 146, in <module>
131
+ eval_model(args)
132
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 79, in eval_model
133
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
134
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
135
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
136
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
137
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
138
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
139
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
140
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
141
+ resolved_config_file = cached_file(
142
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
143
+ raise EnvironmentError(
144
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
145
+ Traceback (most recent call last):
146
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/convert_answer_to_mme.py", line 52, in <module>
147
+ answers = [json.loads(line) for line in open(os.path.join('answers', f'{experiment}.jsonl'))]
148
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation-mask_applied.jsonl'
149
+ =========== Perception ===========
150
+ Traceback (most recent call last):
151
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 164, in <module>
152
+ cal.process_result(results_dir)
153
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 98, in process_result
154
+ lines = open(task_txt, 'r').readlines()
155
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation-mask_applied/existence.txt'
156
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
157
+ import pynvml # type: ignore[import]
158
+ [2025-10-12 17:41:19,450] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
159
+ Traceback (most recent call last):
160
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
161
+ resolved_file = hf_hub_download(
162
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
163
+ validate_repo_id(arg_value)
164
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
165
+ raise HFValidationError(
166
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
167
+
168
+ The above exception was the direct cause of the following exception:
169
+
170
+ Traceback (most recent call last):
171
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
172
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
173
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
174
+ resolved_config_file = cached_file(
175
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
176
+ raise EnvironmentError(
177
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
178
+
179
+ During handling of the above exception, another exception occurred:
180
+
181
+ Traceback (most recent call last):
182
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
183
+ resolved_file = hf_hub_download(
184
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
185
+ validate_repo_id(arg_value)
186
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
187
+ raise HFValidationError(
188
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
189
+
190
+ The above exception was the direct cause of the following exception:
191
+
192
+ Traceback (most recent call last):
193
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
194
+ return _run_code(code, main_globals, None,
195
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
196
+ exec(code, run_globals)
197
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 108, in <module>
198
+ eval_model(args)
199
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 31, in eval_model
200
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
201
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
202
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
203
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
204
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
205
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
206
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
207
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
208
+ resolved_config_file = cached_file(
209
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
210
+ raise EnvironmentError(
211
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
212
+ Traceback (most recent call last):
213
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/eval_science_qa.py", line 45, in <module>
214
+ predictions = [json.loads(line) for line in open(args.result_file)]
215
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/scienceqa/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation-mask_applied.jsonl'
216
+ ==== EXPERIMENT COMPLETED: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation ====
217
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation_20251012_173957.log
218
+ Timestamp: 2025-10-12 17:41:22
219
+ =====================================
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_1e-2_connector-3.0_0.5_1e-2_ablation_20251012_174122.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation_20251012_180629.log ADDED
@@ -0,0 +1,219 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==== STARTING EXPERIMENT: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation ====
2
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation_20251012_180629.log
3
+ Timestamp: 2025-10-12 18:06:30
4
+ =====================================
5
+ Processing: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation
6
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
7
+ import pynvml # type: ignore[import]
8
+ [2025-10-12 18:06:32,829] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ Traceback (most recent call last):
10
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 488, in <module>
11
+ main()
12
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 123, in main
13
+ config_mask = TinyLlavaConfig.from_pretrained(model_args.mask_model_name_or_path)
14
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
15
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
16
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
17
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
18
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
19
+ resolved_config_file = cached_file(
20
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 369, in cached_file
21
+ raise EnvironmentError(
22
+ OSError: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation does not appear to have a file named config.json. Checkout 'https://huggingface.co//nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/tree/main' for available files.
23
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
24
+ import pynvml # type: ignore[import]
25
+ [2025-10-12 18:06:39,994] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
26
+ Traceback (most recent call last):
27
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
28
+ resolved_file = hf_hub_download(
29
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
30
+ validate_repo_id(arg_value)
31
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
32
+ raise HFValidationError(
33
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
34
+
35
+ The above exception was the direct cause of the following exception:
36
+
37
+ Traceback (most recent call last):
38
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
39
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
40
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
41
+ resolved_config_file = cached_file(
42
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
43
+ raise EnvironmentError(
44
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
45
+
46
+ During handling of the above exception, another exception occurred:
47
+
48
+ Traceback (most recent call last):
49
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
50
+ resolved_file = hf_hub_download(
51
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
52
+ validate_repo_id(arg_value)
53
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
54
+ raise HFValidationError(
55
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
56
+
57
+ The above exception was the direct cause of the following exception:
58
+
59
+ Traceback (most recent call last):
60
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
61
+ return _run_code(code, main_globals, None,
62
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
63
+ exec(code, run_globals)
64
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 180, in <module>
65
+ eval_model(args)
66
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 88, in eval_model
67
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
68
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
69
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
70
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
71
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
72
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
73
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
74
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
75
+ resolved_config_file = cached_file(
76
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
77
+ raise EnvironmentError(
78
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
79
+ Traceback (most recent call last):
80
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 31, in <module>
81
+ eval_model(args)
82
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 7, in eval_model
83
+ answers = [json.loads(q) for q in open(os.path.expanduser(args.answers_file), "r")]
84
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation-mask_applied.jsonl'
85
+ Traceback (most recent call last):
86
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MMMU/eval/main_eval_only.py", line 19, in <module>
87
+ output_dict = json.load(open(args.output_path))
88
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation-mask_applied_output.json'
89
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
90
+ import pynvml # type: ignore[import]
91
+ [2025-10-12 18:06:46,191] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
92
+ Traceback (most recent call last):
93
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
94
+ resolved_file = hf_hub_download(
95
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
96
+ validate_repo_id(arg_value)
97
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
98
+ raise HFValidationError(
99
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
100
+
101
+ The above exception was the direct cause of the following exception:
102
+
103
+ Traceback (most recent call last):
104
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
105
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
106
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
107
+ resolved_config_file = cached_file(
108
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
109
+ raise EnvironmentError(
110
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
111
+
112
+ During handling of the above exception, another exception occurred:
113
+
114
+ Traceback (most recent call last):
115
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
116
+ resolved_file = hf_hub_download(
117
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
118
+ validate_repo_id(arg_value)
119
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
120
+ raise HFValidationError(
121
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
122
+
123
+ The above exception was the direct cause of the following exception:
124
+
125
+ Traceback (most recent call last):
126
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
127
+ return _run_code(code, main_globals, None,
128
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
129
+ exec(code, run_globals)
130
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 146, in <module>
131
+ eval_model(args)
132
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 79, in eval_model
133
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
134
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
135
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
136
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
137
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
138
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
139
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
140
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
141
+ resolved_config_file = cached_file(
142
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
143
+ raise EnvironmentError(
144
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
145
+ Traceback (most recent call last):
146
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/convert_answer_to_mme.py", line 52, in <module>
147
+ answers = [json.loads(line) for line in open(os.path.join('answers', f'{experiment}.jsonl'))]
148
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation-mask_applied.jsonl'
149
+ =========== Perception ===========
150
+ Traceback (most recent call last):
151
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 164, in <module>
152
+ cal.process_result(results_dir)
153
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 98, in process_result
154
+ lines = open(task_txt, 'r').readlines()
155
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation-mask_applied/existence.txt'
156
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
157
+ import pynvml # type: ignore[import]
158
+ [2025-10-12 18:07:51,101] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
159
+ Traceback (most recent call last):
160
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
161
+ resolved_file = hf_hub_download(
162
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
163
+ validate_repo_id(arg_value)
164
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
165
+ raise HFValidationError(
166
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
167
+
168
+ The above exception was the direct cause of the following exception:
169
+
170
+ Traceback (most recent call last):
171
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
172
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
173
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
174
+ resolved_config_file = cached_file(
175
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
176
+ raise EnvironmentError(
177
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
178
+
179
+ During handling of the above exception, another exception occurred:
180
+
181
+ Traceback (most recent call last):
182
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
183
+ resolved_file = hf_hub_download(
184
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
185
+ validate_repo_id(arg_value)
186
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
187
+ raise HFValidationError(
188
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
189
+
190
+ The above exception was the direct cause of the following exception:
191
+
192
+ Traceback (most recent call last):
193
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
194
+ return _run_code(code, main_globals, None,
195
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
196
+ exec(code, run_globals)
197
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 108, in <module>
198
+ eval_model(args)
199
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 31, in eval_model
200
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
201
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
202
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
203
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
204
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
205
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
206
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
207
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
208
+ resolved_config_file = cached_file(
209
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
210
+ raise EnvironmentError(
211
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
212
+ Traceback (most recent call last):
213
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/eval_science_qa.py", line 45, in <module>
214
+ predictions = [json.loads(line) for line in open(args.result_file)]
215
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/scienceqa/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation-mask_applied.jsonl'
216
+ ==== EXPERIMENT COMPLETED: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation ====
217
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation_20251012_180629.log
218
+ Timestamp: 2025-10-12 18:07:54
219
+ =====================================
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_3e-2_connector-3.0_0.5_3e-2_ablation_20251012_180754.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-1_connector-3.0_0.5_5e-1_ablation_20251012_183317.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation_20251012_190124.log ADDED
@@ -0,0 +1,219 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==== STARTING EXPERIMENT: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation ====
2
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation_20251012_190124.log
3
+ Timestamp: 2025-10-12 19:01:24
4
+ =====================================
5
+ Processing: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation
6
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
7
+ import pynvml # type: ignore[import]
8
+ [2025-10-12 19:01:27,172] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ Traceback (most recent call last):
10
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 488, in <module>
11
+ main()
12
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 123, in main
13
+ config_mask = TinyLlavaConfig.from_pretrained(model_args.mask_model_name_or_path)
14
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
15
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
16
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
17
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
18
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
19
+ resolved_config_file = cached_file(
20
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 369, in cached_file
21
+ raise EnvironmentError(
22
+ OSError: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation does not appear to have a file named config.json. Checkout 'https://huggingface.co//nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/tree/main' for available files.
23
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
24
+ import pynvml # type: ignore[import]
25
+ [2025-10-12 19:01:34,395] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
26
+ Traceback (most recent call last):
27
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
28
+ resolved_file = hf_hub_download(
29
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
30
+ validate_repo_id(arg_value)
31
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
32
+ raise HFValidationError(
33
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Use `repo_type` argument if needed.
34
+
35
+ The above exception was the direct cause of the following exception:
36
+
37
+ Traceback (most recent call last):
38
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
39
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
40
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
41
+ resolved_config_file = cached_file(
42
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
43
+ raise EnvironmentError(
44
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
45
+
46
+ During handling of the above exception, another exception occurred:
47
+
48
+ Traceback (most recent call last):
49
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
50
+ resolved_file = hf_hub_download(
51
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
52
+ validate_repo_id(arg_value)
53
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
54
+ raise HFValidationError(
55
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Use `repo_type` argument if needed.
56
+
57
+ The above exception was the direct cause of the following exception:
58
+
59
+ Traceback (most recent call last):
60
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
61
+ return _run_code(code, main_globals, None,
62
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
63
+ exec(code, run_globals)
64
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 180, in <module>
65
+ eval_model(args)
66
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 88, in eval_model
67
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
68
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
69
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
70
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
71
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
72
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
73
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
74
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
75
+ resolved_config_file = cached_file(
76
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
77
+ raise EnvironmentError(
78
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
79
+ Traceback (most recent call last):
80
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 31, in <module>
81
+ eval_model(args)
82
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 7, in eval_model
83
+ answers = [json.loads(q) for q in open(os.path.expanduser(args.answers_file), "r")]
84
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation-mask_applied.jsonl'
85
+ Traceback (most recent call last):
86
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MMMU/eval/main_eval_only.py", line 19, in <module>
87
+ output_dict = json.load(open(args.output_path))
88
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation-mask_applied_output.json'
89
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
90
+ import pynvml # type: ignore[import]
91
+ [2025-10-12 19:01:40,892] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
92
+ Traceback (most recent call last):
93
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
94
+ resolved_file = hf_hub_download(
95
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
96
+ validate_repo_id(arg_value)
97
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
98
+ raise HFValidationError(
99
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Use `repo_type` argument if needed.
100
+
101
+ The above exception was the direct cause of the following exception:
102
+
103
+ Traceback (most recent call last):
104
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
105
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
106
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
107
+ resolved_config_file = cached_file(
108
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
109
+ raise EnvironmentError(
110
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
111
+
112
+ During handling of the above exception, another exception occurred:
113
+
114
+ Traceback (most recent call last):
115
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
116
+ resolved_file = hf_hub_download(
117
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
118
+ validate_repo_id(arg_value)
119
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
120
+ raise HFValidationError(
121
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Use `repo_type` argument if needed.
122
+
123
+ The above exception was the direct cause of the following exception:
124
+
125
+ Traceback (most recent call last):
126
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
127
+ return _run_code(code, main_globals, None,
128
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
129
+ exec(code, run_globals)
130
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 146, in <module>
131
+ eval_model(args)
132
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 79, in eval_model
133
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
134
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
135
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
136
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
137
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
138
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
139
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
140
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
141
+ resolved_config_file = cached_file(
142
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
143
+ raise EnvironmentError(
144
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
145
+ Traceback (most recent call last):
146
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/convert_answer_to_mme.py", line 52, in <module>
147
+ answers = [json.loads(line) for line in open(os.path.join('answers', f'{experiment}.jsonl'))]
148
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation-mask_applied.jsonl'
149
+ =========== Perception ===========
150
+ Traceback (most recent call last):
151
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 164, in <module>
152
+ cal.process_result(results_dir)
153
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 98, in process_result
154
+ lines = open(task_txt, 'r').readlines()
155
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation-mask_applied/existence.txt'
156
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
157
+ import pynvml # type: ignore[import]
158
+ [2025-10-12 19:02:46,283] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
159
+ Traceback (most recent call last):
160
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
161
+ resolved_file = hf_hub_download(
162
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
163
+ validate_repo_id(arg_value)
164
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
165
+ raise HFValidationError(
166
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Use `repo_type` argument if needed.
167
+
168
+ The above exception was the direct cause of the following exception:
169
+
170
+ Traceback (most recent call last):
171
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
172
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
173
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
174
+ resolved_config_file = cached_file(
175
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
176
+ raise EnvironmentError(
177
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
178
+
179
+ During handling of the above exception, another exception occurred:
180
+
181
+ Traceback (most recent call last):
182
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
183
+ resolved_file = hf_hub_download(
184
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
185
+ validate_repo_id(arg_value)
186
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
187
+ raise HFValidationError(
188
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Use `repo_type` argument if needed.
189
+
190
+ The above exception was the direct cause of the following exception:
191
+
192
+ Traceback (most recent call last):
193
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
194
+ return _run_code(code, main_globals, None,
195
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
196
+ exec(code, run_globals)
197
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 108, in <module>
198
+ eval_model(args)
199
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 31, in eval_model
200
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
201
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
202
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
203
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
204
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
205
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
206
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
207
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
208
+ resolved_config_file = cached_file(
209
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
210
+ raise EnvironmentError(
211
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
212
+ Traceback (most recent call last):
213
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/eval_science_qa.py", line 45, in <module>
214
+ predictions = [json.loads(line) for line in open(args.result_file)]
215
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/scienceqa/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation-mask_applied.jsonl'
216
+ ==== EXPERIMENT COMPLETED: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation ====
217
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation_20251012_190124.log
218
+ Timestamp: 2025-10-12 19:02:49
219
+ =====================================
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-1_connector-3.0_0.5_7e-1_ablation_20251012_190249.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-2_connector-3.0_0.5_7e-2_ablation_20251012_192708.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-1_connector-3.0_0.5_9e-1_ablation_20251012_195147.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-2_connector-3.0_0.5_9e-2_ablation_20251012_202158.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.7_2e-1_connector-3.0_0.7_2e-1_ablation_20251012_204649.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.9_2e-1_connector-3.0_0.9_2e-1_ablation_20251012_211112.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.1_2e-1_connector-3.0_1.1_2e-1_ablation_20251012_213606.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation_20251012_220114.log ADDED
@@ -0,0 +1,219 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==== STARTING EXPERIMENT: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation ====
2
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation_20251012_220114.log
3
+ Timestamp: 2025-10-12 22:01:14
4
+ =====================================
5
+ Processing: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation
6
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
7
+ import pynvml # type: ignore[import]
8
+ [2025-10-12 22:01:17,821] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ Traceback (most recent call last):
10
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 488, in <module>
11
+ main()
12
+ File "/nfs/ywang29/TinyLLaVA/scripts/apply_masks.py", line 123, in main
13
+ config_mask = TinyLlavaConfig.from_pretrained(model_args.mask_model_name_or_path)
14
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
15
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
16
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
17
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
18
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
19
+ resolved_config_file = cached_file(
20
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 369, in cached_file
21
+ raise EnvironmentError(
22
+ OSError: /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation does not appear to have a file named config.json. Checkout 'https://huggingface.co//nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/tree/main' for available files.
23
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
24
+ import pynvml # type: ignore[import]
25
+ [2025-10-12 22:01:25,210] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
26
+ Traceback (most recent call last):
27
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
28
+ resolved_file = hf_hub_download(
29
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
30
+ validate_repo_id(arg_value)
31
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
32
+ raise HFValidationError(
33
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
34
+
35
+ The above exception was the direct cause of the following exception:
36
+
37
+ Traceback (most recent call last):
38
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
39
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
40
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
41
+ resolved_config_file = cached_file(
42
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
43
+ raise EnvironmentError(
44
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
45
+
46
+ During handling of the above exception, another exception occurred:
47
+
48
+ Traceback (most recent call last):
49
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
50
+ resolved_file = hf_hub_download(
51
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
52
+ validate_repo_id(arg_value)
53
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
54
+ raise HFValidationError(
55
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
56
+
57
+ The above exception was the direct cause of the following exception:
58
+
59
+ Traceback (most recent call last):
60
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
61
+ return _run_code(code, main_globals, None,
62
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
63
+ exec(code, run_globals)
64
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 180, in <module>
65
+ eval_model(args)
66
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_mmmu.py", line 88, in eval_model
67
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
68
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
69
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
70
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
71
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
72
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
73
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
74
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
75
+ resolved_config_file = cached_file(
76
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
77
+ raise EnvironmentError(
78
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
79
+ Traceback (most recent call last):
80
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 31, in <module>
81
+ eval_model(args)
82
+ File "/nfs/ywang29/TinyLLaVA/scripts/convert_answer_to_mmmu.py", line 7, in eval_model
83
+ answers = [json.loads(q) for q in open(os.path.expanduser(args.answers_file), "r")]
84
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation-mask_applied.jsonl'
85
+ Traceback (most recent call last):
86
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MMMU/eval/main_eval_only.py", line 19, in <module>
87
+ output_dict = json.load(open(args.output_path))
88
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/MMMU/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation-mask_applied_output.json'
89
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
90
+ import pynvml # type: ignore[import]
91
+ [2025-10-12 22:01:31,630] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
92
+ Traceback (most recent call last):
93
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
94
+ resolved_file = hf_hub_download(
95
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
96
+ validate_repo_id(arg_value)
97
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
98
+ raise HFValidationError(
99
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
100
+
101
+ The above exception was the direct cause of the following exception:
102
+
103
+ Traceback (most recent call last):
104
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
105
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
106
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
107
+ resolved_config_file = cached_file(
108
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
109
+ raise EnvironmentError(
110
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
111
+
112
+ During handling of the above exception, another exception occurred:
113
+
114
+ Traceback (most recent call last):
115
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
116
+ resolved_file = hf_hub_download(
117
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
118
+ validate_repo_id(arg_value)
119
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
120
+ raise HFValidationError(
121
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
122
+
123
+ The above exception was the direct cause of the following exception:
124
+
125
+ Traceback (most recent call last):
126
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
127
+ return _run_code(code, main_globals, None,
128
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
129
+ exec(code, run_globals)
130
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 146, in <module>
131
+ eval_model(args)
132
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_loader.py", line 79, in eval_model
133
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
134
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
135
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
136
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
137
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
138
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
139
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
140
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
141
+ resolved_config_file = cached_file(
142
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
143
+ raise EnvironmentError(
144
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
145
+ Traceback (most recent call last):
146
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/convert_answer_to_mme.py", line 52, in <module>
147
+ answers = [json.loads(line) for line in open(os.path.join('answers', f'{experiment}.jsonl'))]
148
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation-mask_applied.jsonl'
149
+ =========== Perception ===========
150
+ Traceback (most recent call last):
151
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 164, in <module>
152
+ cal.process_result(results_dir)
153
+ File "/s3-code/ywang29/datasets/tinyllava/eval/MME/eval_tool/calculation.py", line 98, in process_result
154
+ lines = open(task_txt, 'r').readlines()
155
+ FileNotFoundError: [Errno 2] No such file or directory: 'answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation-mask_applied/existence.txt'
156
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
157
+ import pynvml # type: ignore[import]
158
+ [2025-10-12 22:02:36,588] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
159
+ Traceback (most recent call last):
160
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
161
+ resolved_file = hf_hub_download(
162
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
163
+ validate_repo_id(arg_value)
164
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
165
+ raise HFValidationError(
166
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
167
+
168
+ The above exception was the direct cause of the following exception:
169
+
170
+ Traceback (most recent call last):
171
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 38, in load_pretrained_model
172
+ model = TinyLlavaForConditionalGeneration.from_pretrained(model_name_or_path,low_cpu_mem_usage=True)
173
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/modeling_utils.py", line 3015, in from_pretrained
174
+ resolved_config_file = cached_file(
175
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
176
+ raise EnvironmentError(
177
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
178
+
179
+ During handling of the above exception, another exception occurred:
180
+
181
+ Traceback (most recent call last):
182
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 398, in cached_file
183
+ resolved_file = hf_hub_download(
184
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn
185
+ validate_repo_id(arg_value)
186
+ File "/opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id
187
+ raise HFValidationError(
188
+ huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Use `repo_type` argument if needed.
189
+
190
+ The above exception was the direct cause of the following exception:
191
+
192
+ Traceback (most recent call last):
193
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
194
+ return _run_code(code, main_globals, None,
195
+ File "/opt/conda/envs/tinyllava/lib/python3.10/runpy.py", line 86, in _run_code
196
+ exec(code, run_globals)
197
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 108, in <module>
198
+ eval_model(args)
199
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/model_vqa_science.py", line 31, in eval_model
200
+ model, tokenizer, image_processor, context_len = load_pretrained_model(model_path)
201
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/model/load_model.py", line 40, in load_pretrained_model
202
+ model_config = TinyLlavaConfig.from_pretrained(model_name_or_path)
203
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 602, in from_pretrained
204
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
205
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 631, in get_config_dict
206
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
207
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/configuration_utils.py", line 686, in _get_config_dict
208
+ resolved_config_file = cached_file(
209
+ File "/nfs/ywang29/TinyLLaVA/transformers/src/transformers/utils/hub.py", line 462, in cached_file
210
+ raise EnvironmentError(
211
+ OSError: Incorrect path_or_model_id: '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation/mask_applied'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
212
+ Traceback (most recent call last):
213
+ File "/nfs/ywang29/TinyLLaVA/tinyllava/eval/eval_science_qa.py", line 45, in <module>
214
+ predictions = [json.loads(line) for line in open(args.result_file)]
215
+ FileNotFoundError: [Errno 2] No such file or directory: '/s3-code/ywang29/datasets/tinyllava/eval/scienceqa/answers/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation-mask_applied.jsonl'
216
+ ==== EXPERIMENT COMPLETED: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation ====
217
+ Log File: eval_qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation_20251012_220114.log
218
+ Timestamp: 2025-10-12 22:02:39
219
+ =====================================
logs_oct11/logs_oct10.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49b2a4f137f8b5451c7553bd7aff99979ef50535267bd54c9681bcbc8c342cde
3
+ size 917335
logs_oct11/logs_oct9.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88c8911009bc76194a78b1b78dc3bebabc8069034e88401141577ac3aac0e6e4
3
+ size 471303
logs_oct11/plot_ablation.py ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+ import matplotlib
3
+ import matplotlib.pyplot as plt
4
+ from mpl_toolkits.mplot3d import Axes3D
5
+
6
+ # 使用Agg后端(无GUI)
7
+ matplotlib.use("Agg")
8
+
9
+ # 定义x轴和y轴的数据
10
+ x_values = np.array([1, 3, 5, 7, 9]) # init value
11
+ y_values = np.array([0.1, 0.3, 0.5, 0.7, 0.9, 1.1, 1.3, 1.5, 1.7, 1.9, 2.1, 2.3, 2.5, 2.7, 2.9]) # temperature
12
+
13
+ # 创建网格
14
+ X, Y = np.meshgrid(x_values, y_values)
15
+
16
+ # Z轴数据(performance)- 这里使用示例数据
17
+ # 请将这里替换为你的实际数据,应该是一个15x5的数组
18
+ # 示例:使用一个简单的函数生成数据
19
+ # Z = np.sin(np.sqrt(X**2 + Y**2)) * 100 + np.random.randn(15, 5) * 5
20
+
21
+ # 如果你有实际的performance数据,请按照以下格式替换:
22
+ Z = np.array([
23
+ [11.9, 14.8, 13.7, 13.0, 10.8], # temperature=0.1时,对应5个init value的performance
24
+ [17.5, 10.8, 10.3, 14.3, 13.0], # temperature=0.3时,对应5个init value的performance
25
+ [29.2, 31.7, 13.5, 10.1, 14.8], # temperature=0.5时,对应5个init value的performance
26
+ [20.0, 30.8, 23.3, 14.2, 11.1], # temperature=0.7时,对应5个init value的performance
27
+ [10.9, 23.3, 25.0, 24.2, 10.9], # temperature=0.9时,对应5个init value的performance
28
+ [23.3, 24.2, 23.3, 29.2, 21.7], # temperature=1.1时,对应5个init value的performance
29
+ [21.7, 27.5, 22.5, 25.8, 25.0], # temperature=1.3时,对应5个init value的performance
30
+ [11.5, 23.3, 17.5, 26.7, 27.5], # temperature=1.5时,对应5个init value的performance
31
+ [24.2, 24.2, 16.7, 23.3, 29.2], # temperature=1.7时,对应5个init value的performance
32
+ [18.3, 24.2, 27.5, 20.0, 25.8], # temperature=1.9时,对应5个init value的performance
33
+ [25.8, 25.8, 26.7, 21.7, 28.3], # temperature=2.1时,对应5个init value的performance
34
+ [28.3, 26.2, 23.3, 20.8, 26.7], # temperature=2.3时,对应5个init value的performance
35
+ [20.0, 24.2, 20.0, 20.0, 28.3], # temperature=2.5时,对应5个init value的performance
36
+ [23.3, 26.2, 21.7, 24.2, 24.2], # temperature=2.7时,对应5个init value的performance
37
+ [15.8, 22.5, 20.8, 25.8, 23.3], # temperature=2.9时,对应5个init value的performance
38
+ ])
39
+
40
+ # 创建3D图形
41
+ fig = plt.figure(figsize=(12, 8))
42
+ ax = fig.add_subplot(111, projection='3d')
43
+
44
+ # 绘制3D表面图(不显示散点)
45
+ surf = ax.plot_surface(X, Y, Z, cmap='RdYlBu_r', alpha=0.8, edgecolor='none')
46
+
47
+ # 设置轴标签
48
+ ax.set_xlabel('Init Value', fontsize=12, labelpad=10)
49
+ ax.set_ylabel('Temperature', fontsize=12, labelpad=10)
50
+ ax.set_zlabel('Performance', fontsize=12, labelpad=10)
51
+
52
+ # 设置标题(字体更大,间隙更小)
53
+ ax.set_title('Performance vs Init Value and Temperature', fontsize=18, pad=10)
54
+
55
+ # 添加颜色条(更窄,离图更近)
56
+ colorbar = fig.colorbar(surf, ax=ax, shrink=0.5, aspect=10, pad=0.05)
57
+ colorbar.set_label('Performance', rotation=270, labelpad=15)
58
+
59
+ # 设置视角(可以调整以获得最佳视图)
60
+ ax.view_init(elev=30, azim=45)
61
+
62
+ # 明确设置x轴和y轴的刻度
63
+ ax.set_xticks(x_values)
64
+ ax.set_yticks(y_values)
65
+ # 如果需要,也可以设置z轴的刻度
66
+ # ax.set_zticks(np.linspace(Z.min(), Z.max(), 5))
67
+
68
+ # 移除网格线
69
+ ax.grid(False)
70
+
71
+ # 设置透明背景
72
+ ax.xaxis.pane.fill = False
73
+ ax.yaxis.pane.fill = False
74
+ ax.zaxis.pane.fill = False
75
+ ax.xaxis.pane.set_edgecolor('none')
76
+ ax.yaxis.pane.set_edgecolor('none')
77
+ ax.zaxis.pane.set_edgecolor('none')
78
+
79
+ # 保存3D图(透明背景)
80
+ plt.tight_layout()
81
+ plt.savefig('3d_surface_plot.png', dpi=300, bbox_inches='tight', transparent=True)
82
+ plt.close()
83
+
84
+ print("3D表面图已保存为: 3d_surface_plot.png")
85
+
86
+ # 创建等高线图
87
+ plt.figure(figsize=(7, 5))
88
+
89
+ num_levels = 15
90
+ # 使用 'RdYlBu_r' 颜色映射
91
+ contourf = plt.contourf(X, Y, Z, levels=num_levels, cmap='RdYlBu_r', alpha=0.7)
92
+
93
+ # 加粗的黑色等高线,不贴数值标签
94
+ contour_lines = plt.contour(X, Y, Z, levels=num_levels, colors='black', alpha=0.5, linewidths=1.5)
95
+
96
+ # 添加颜色条(更窄,离图更近)
97
+ cbar = plt.colorbar(contourf, shrink=0.8, aspect=15, pad=0.02)
98
+ cbar.set_label("Performance", rotation=270, labelpad=15)
99
+
100
+ # 设置轴标签
101
+ plt.xlabel('Init Value', fontsize=12)
102
+ plt.ylabel('Temperature', fontsize=12)
103
+
104
+ # 设置标题(字体更大,间隙更小)
105
+ plt.title('Performance Contour Plot', fontsize=18, pad=10)
106
+
107
+ # 移除网格
108
+ plt.grid(False)
109
+
110
+ # 显示x和y轴的刻度
111
+ plt.xticks(x_values)
112
+ plt.yticks(y_values)
113
+
114
+ # 保存等高线图(透明背景)
115
+ plt.tight_layout()
116
+ plt.savefig('contour_plot.png', dpi=300, bbox_inches='tight', transparent=True)
117
+ plt.close()
118
+
119
+ print("等高线图已保存为: contour_plot.png")
120
+ print("所有图片已成功保存!")
logs_oct11/pretrain_qwen.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:420378c2943468d547bf10619389107e3e52783ba88768e51e07099df45eadc6
3
+ size 1473310868
logs_oct11/pyproject.toml ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [build-system]
2
+ requires = ["setuptools>=61.0"]
3
+ build-backend = "setuptools.build_meta"
4
+
5
+ [project]
6
+ name = "tinyllava"
7
+ version = "1.0.0"
8
+ description = "A Framework of Small-scale Large Multimodal Models."
9
+ readme = "README.md"
10
+ requires-python = ">=3.9"
11
+ classifiers = [
12
+ "Programming Language :: Python :: 3",
13
+ "License :: OSI Approved :: Apache Software License",
14
+ ]
15
+ dependencies = [
16
+ "torch==2.1.0", "torchvision==0.16.0", "tiktoken", "openpyxl", "tensorboardX",
17
+ "transformers==4.40.1", "tokenizers==0.19.0", "sentencepiece==0.1.99", "shortuuid",
18
+ "accelerate==0.27.2", "bitsandbytes==0.41.0", "peft==0.10.0",
19
+ "pydantic<2,>=1", "markdown2[all]", "numpy==1.26.4", "scikit-learn==1.2.2",
20
+ "gradio==3.35.2", "gradio_client==0.2.9",
21
+ "requests", "httpx==0.24.0", "uvicorn", "fastapi",
22
+ "einops==0.6.1", "einops-exts==0.0.4", "timm==0.6.13",
23
+ "deepspeed==0.14.0", "ninja", "wandb",
24
+ ]
25
+
26
+ [project.optional-dependencies]
27
+ train = ["deepspeed==0.14.0", "ninja", "wandb"]
28
+
29
+ [project.urls]
30
+ "Homepage" = "https://github.com/DLCV-BUAA/TinyLLaVABench"
31
+ "Bug Tracker" = "https://github.com/DLCV-BUAA/TinyLLaVABench/issues"
32
+
33
+ [tool.setuptools.packages.find]
34
+ exclude = ["assets*", "benchmark*", "docs", "dist*", "playground*", "scripts*", "tests*"]
35
+
36
+ [tool.wheel]
37
+ exclude = ["assets*", "benchmark*", "docs", "dist*", "playground*", "scripts*", "tests*"]
38
+
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation_20251012_052820.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation_20251012_155158.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ ==== STARTING EXPERIMENT: qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation ====
2
+ Log File: qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_1.9_2e-1_connector-1.0_1.9_2e-1_ablation_20251012_155158.log
3
+ Timestamp: 2025-10-12 15:51:58
4
+ =====================================
5
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
6
+ import pynvml # type: ignore[import]
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.1_2e-1_connector-1.0_2.1_2e-1_ablation_20251012_060227.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.3_2e-1_connector-1.0_2.3_2e-1_ablation_20251012_081931.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.5_2e-1_connector-1.0_2.5_2e-1_ablation_20251012_090245.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.7_2e-1_connector-1.0_2.7_2e-1_ablation_20251012_100457.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-1.0_2.9_2e-1_connector-1.0_2.9_2e-1_ablation_20251012_113628.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_1e-2_connector-3.0_0.5_1e-2_ablation_20251012_091538.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_2e-1_connector-3.0_0.5_2e-1_ablation_20251012_123546.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_3e-2_connector-3.0_0.5_3e-2_ablation_20251012_100541.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-1_connector-3.0_0.5_5e-1_ablation_20251012_052751.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_5e-2_connector-3.0_0.5_5e-2_ablation_20251012_113222.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-1_connector-3.0_0.5_7e-1_ablation_20251012_071236.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_7e-2_connector-3.0_0.5_7e-2_ablation_20251012_132440.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-1_connector-3.0_0.5_9e-1_ablation_20251012_081416.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.5_9e-2_connector-3.0_0.5_9e-2_ablation_20251012_142902.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.7_2e-1_connector-3.0_0.7_2e-1_ablation_20251012_133702.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_0.9_2e-1_connector-3.0_0.9_2e-1_ablation_20251012_144110.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.1_2e-1_connector-3.0_1.1_2e-1_ablation_20251012_151526.log ADDED
The diff for this file is too large to render. See raw diff
 
logs_oct11/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation_20251012_154943.log ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ==== STARTING EXPERIMENT: qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation ====
2
+ Log File: qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation_20251012_154943.log
3
+ Timestamp: 2025-10-12 15:49:43
4
+ =====================================
5
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
6
+ import pynvml # type: ignore[import]
7
+ [2025-10-12 15:49:45,676] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
8
+ [2025-10-12 15:49:48,368] [WARNING] [runner.py:202:fetch_hostfile] Unable to find hostfile, will proceed with training with local resources only.
9
+ [2025-10-12 15:49:48,370] [INFO] [runner.py:568:main] cmd = /opt/conda/envs/tinyllava/bin/python3.10 -u -m deepspeed.launcher.launch --world_info=eyJsb2NhbGhvc3QiOiBbMCwgMSwgMiwgMywgNCwgNSwgNiwgN119 --master_addr=127.0.0.1 --master_port=29501 --enable_each_rank_log=None tinyllava/train/train.py --deepspeed ./scripts/zero3.json --data_path /s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json --image_folder /s3-code/ywang29/datasets/tinyllava --is_multimodal True --conv_version qwen2_base --model_name_or_path Qwen/Qwen2.5-0.5B --vision_tower google/siglip-so400m-patch14-384 --vision_tower2 --connector_type mlp2x_gelu --mm_vision_select_layer -2 --image_aspect_ratio square --attn_implementation flash_attention_2 --bf16 True --training_recipe common --tune_type_llm full --tune_type_vision_tower frozen --tune_vision_tower_from_layer 0 --tune_type_connector full --group_by_modality_length True --pretrained_model_path /nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain --output_dir /nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation --num_train_epochs 1 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --gradient_accumulation_steps 4 --evaluation_strategy no --learning_rate 2e-1 --weight_decay 0. --warmup_ratio 0.03 --lr_scheduler_type cosine --logging_steps 1 --tf32 False --model_max_length 2048 --gradient_checkpointing True --dataloader_num_workers 8 --lazy_preprocess True --report_to tensorboard --tokenizer_use_fast False --run_name tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune --subnet_mode_text both --subnet_type_text None --mask_type_text soft --init_mean_text 3.0 --temperature_attn_text 1.3 --temperature_mlp_text 1.3 --backward_type_text normal --masked_layers_text all --subnet_mode_vision both --subnet_type_vision None --mask_type_vision soft --init_mean_vision 3.0 --temperature_attn_vision 1.3 --temperature_mlp_vision 1.3 --backward_type_vision normal --masked_layers_vision all --subnet_type_connector global --mask_type_connector soft --init_mean_connector 3.0 --temperature_connector 1.3 --backward_type_connector normal --mm_projector_lr 2e-1 --seed 42 --mask_model llm-connector --save_strategy steps --save_steps 50000 --save_total_limit 1 --train_data_ratio 0.1
10
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
11
+ import pynvml # type: ignore[import]
12
+ [2025-10-12 15:49:50,940] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
13
+ [2025-10-12 15:49:52,009] [INFO] [launch.py:138:main] 0 NCCL_VERSION=2.21.5
14
+ [2025-10-12 15:49:52,009] [INFO] [launch.py:138:main] 0 NCCL_SOCKET_IFNAME=eth
15
+ [2025-10-12 15:49:52,009] [INFO] [launch.py:145:main] WORLD INFO DICT: {'localhost': [0, 1, 2, 3, 4, 5, 6, 7]}
16
+ [2025-10-12 15:49:52,009] [INFO] [launch.py:151:main] nnodes=1, num_local_procs=8, node_rank=0
17
+ [2025-10-12 15:49:52,009] [INFO] [launch.py:162:main] global_rank_mapping=defaultdict(<class 'list'>, {'localhost': [0, 1, 2, 3, 4, 5, 6, 7]})
18
+ [2025-10-12 15:49:52,009] [INFO] [launch.py:163:main] dist_world_size=8
19
+ [2025-10-12 15:49:52,009] [INFO] [launch.py:165:main] Setting CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
20
+ [2025-10-12 15:49:52,012] [INFO] [launch.py:253:main] process 445249 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=0', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
21
+ [2025-10-12 15:49:52,014] [INFO] [launch.py:253:main] process 445250 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=1', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
22
+ [2025-10-12 15:49:52,016] [INFO] [launch.py:253:main] process 445251 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=2', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
23
+ [2025-10-12 15:49:52,018] [INFO] [launch.py:253:main] process 445252 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=3', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
24
+ [2025-10-12 15:49:52,021] [INFO] [launch.py:253:main] process 445253 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=4', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
25
+ [2025-10-12 15:49:52,023] [INFO] [launch.py:253:main] process 445254 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=5', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
26
+ [2025-10-12 15:49:52,025] [INFO] [launch.py:253:main] process 445255 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=6', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
27
+ [2025-10-12 15:49:52,027] [INFO] [launch.py:253:main] process 445256 spawned with command: ['/opt/conda/envs/tinyllava/bin/python3.10', '-u', 'tinyllava/train/train.py', '--local_rank=7', '--deepspeed', './scripts/zero3.json', '--data_path', '/s3-code/ywang29/datasets/tinyllava/text_files/llava_v1_5_mix665k.json', '--image_folder', '/s3-code/ywang29/datasets/tinyllava', '--is_multimodal', 'True', '--conv_version', 'qwen2_base', '--model_name_or_path', 'Qwen/Qwen2.5-0.5B', '--vision_tower', 'google/siglip-so400m-patch14-384', '--vision_tower2', '', '--connector_type', 'mlp2x_gelu', '--mm_vision_select_layer', '-2', '--image_aspect_ratio', 'square', '--attn_implementation', 'flash_attention_2', '--bf16', 'True', '--training_recipe', 'common', '--tune_type_llm', 'full', '--tune_type_vision_tower', 'frozen', '--tune_vision_tower_from_layer', '0', '--tune_type_connector', 'full', '--group_by_modality_length', 'True', '--pretrained_model_path', '/nfs/ywang29/TinyLLaVA/checkpoints/tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-pretrain', '--output_dir', '/nfs/ywang29/TinyLLaVA/checkpoints/qwen2.5-0_5b_base_masktune_42_llm-connector_text-3.0_1.3_2e-1_connector-3.0_1.3_2e-1_ablation', '--num_train_epochs', '1', '--per_device_train_batch_size', '4', '--per_device_eval_batch_size', '4', '--gradient_accumulation_steps', '4', '--evaluation_strategy', 'no', '--learning_rate', '2e-1', '--weight_decay', '0.', '--warmup_ratio', '0.03', '--lr_scheduler_type', 'cosine', '--logging_steps', '1', '--tf32', 'False', '--model_max_length', '2048', '--gradient_checkpointing', 'True', '--dataloader_num_workers', '8', '--lazy_preprocess', 'True', '--report_to', 'tensorboard', '--tokenizer_use_fast', 'False', '--run_name', 'tiny-llava-Qwen2.5-0.5B-siglip-so400m-patch14-384-qwen2.5-0_5b_base-masktune', '--subnet_mode_text', 'both', '--subnet_type_text', 'None', '--mask_type_text', 'soft', '--init_mean_text', '3.0', '--temperature_attn_text', '1.3', '--temperature_mlp_text', '1.3', '--backward_type_text', 'normal', '--masked_layers_text', 'all', '--subnet_mode_vision', 'both', '--subnet_type_vision', 'None', '--mask_type_vision', 'soft', '--init_mean_vision', '3.0', '--temperature_attn_vision', '1.3', '--temperature_mlp_vision', '1.3', '--backward_type_vision', 'normal', '--masked_layers_vision', 'all', '--subnet_type_connector', 'global', '--mask_type_connector', 'soft', '--init_mean_connector', '3.0', '--temperature_connector', '1.3', '--backward_type_connector', 'normal', '--mm_projector_lr', '2e-1', '--seed', '42', '--mask_model', 'llm-connector', '--save_strategy', 'steps', '--save_steps', '50000', '--save_total_limit', '1', '--train_data_ratio', '0.1']
28
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
29
+ import pynvml # type: ignore[import]
30
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
31
+ import pynvml # type: ignore[import]
32
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
33
+ import pynvml # type: ignore[import]
34
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
35
+ import pynvml # type: ignore[import]
36
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
37
+ import pynvml # type: ignore[import]
38
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
39
+ import pynvml # type: ignore[import]
40
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
41
+ import pynvml # type: ignore[import]
42
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/torch/cuda/__init__.py:51: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.
43
+ import pynvml # type: ignore[import]
44
+ [2025-10-12 15:49:58,674] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
45
+ [2025-10-12 15:49:58,716] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
46
+ [2025-10-12 15:49:58,770] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
47
+ [2025-10-12 15:49:58,770] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
48
+ [2025-10-12 15:49:58,780] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
49
+ [2025-10-12 15:49:58,780] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
50
+ [2025-10-12 15:49:58,785] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
51
+ [2025-10-12 15:49:58,787] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator to cuda (auto detect)
52
+ [2025-10-12 15:49:59,093] [INFO] [comm.py:637:init_distributed] cdb=None
53
+ [2025-10-12 15:49:59,093] [INFO] [comm.py:668:init_distributed] Initializing TorchBackend in DeepSpeed with backend nccl
54
+ [2025-10-12 15:49:59,124] [INFO] [comm.py:637:init_distributed] cdb=None
55
+ [2025-10-12 15:49:59,171] [INFO] [comm.py:637:init_distributed] cdb=None
56
+ [2025-10-12 15:49:59,173] [INFO] [comm.py:637:init_distributed] cdb=None
57
+ [2025-10-12 15:49:59,182] [INFO] [comm.py:637:init_distributed] cdb=None
58
+ [2025-10-12 15:49:59,185] [INFO] [comm.py:637:init_distributed] cdb=None
59
+ [2025-10-12 15:49:59,189] [INFO] [comm.py:637:init_distributed] cdb=None
60
+ [2025-10-12 15:49:59,190] [INFO] [comm.py:637:init_distributed] cdb=None
61
+ Apply masks for the following modules: ['llm', 'connector']
62
+ {'llm': {'model_name_or_path': 'Qwen/Qwen2.5-0.5B', 'cache_dir': None, 'attn_implementation': 'flash_attention_2', 'subnet_mode': 'both', 'subnet_type': 'None', 'sparsity_attn': None, 'sparsity_mlp': None, 'threshold_attn': None, 'threshold_mlp': None, 'temperature_attn': 1.3, 'temperature_mlp': 1.3, 'masked_layers': 'all', 'mask_type': 'soft', 'backward_type': 'normal'}, 'vision_tower': {'model_name_or_path': 'google/siglip-so400m-patch14-384'}, 'connector': {'connector_type': 'mlp2x_gelu', 'subnet_type': 'global', 'threshold': None, 'sparsity': None, 'temperature': 1.3, 'mask_type': 'soft', 'backward_type': 'normal'}}
63
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
64
+ warnings.warn(
65
+ Apply masks for the following modules: ['llm', 'connector']
66
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
67
+ warnings.warn(
68
+ Apply masks for the following modules: ['llm', 'connector']
69
+ Apply masks for the following modules: ['llm', 'connector']
70
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
71
+ warnings.warn(
72
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
73
+ warnings.warn(
74
+ Apply masks for the following modules: ['llm', 'connector']
75
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
76
+ warnings.warn(
77
+ Apply masks for the following modules: ['llm', 'connector']
78
+ Apply masks for the following modules: ['llm', 'connector']
79
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
80
+ warnings.warn(
81
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
82
+ warnings.warn(
83
+ Apply masks for the following modules: ['llm', 'connector']
84
+ /opt/conda/envs/tinyllava/lib/python3.10/site-packages/huggingface_hub/file_download.py:945: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
85
+ warnings.warn(
86
+ TinyLlavaConfig {
87
+ "backward_type_connector": "normal",
88
+ "cache_dir": null,
89
+ "connector_type": "mlp2x_gelu",
90
+ "hidden_size": 896,
91
+ "ignore_index": -100,
92
+ "image_aspect_ratio": "square",
93
+ "image_token_index": -200,
94
+ "llm_model_name_or_path": "Qwen/Qwen2.5-0.5B",
95
+ "mask_model": [
96
+ "llm",
97
+ "connector"
98
+ ],
99
+ "mask_type_connector": "soft",
100
+ "model_type": "tinyllava",
101
+ "num_queries": 128,
102
+ "num_resampler_layers": 3,
103
+ "pad_token": null,
104
+ "resampler_hidden_size": 768,
105
+ "sparsity_connector": null,
106
+ "subnet_type_connector": "global",
107
+ "temperature_connector": 1.3,
108
+ "text_config": {
109
+ "_name_or_path": "Qwen/Qwen2.5-0.5B",
110
+ "architectures": [
111
+ "Qwen2ForCausalLM"
112
+ ],
113
+ "backward_type": "normal",
114
+ "bos_token_id": 151643,
115
+ "eos_token_id": 151643,
116
+ "hidden_size": 896,
117
+ "intermediate_size": 4864,
118
+ "mask_type": "soft",
119
+ "masked_layers": "all",
120
+ "max_position_embeddings": 32768,
121
+ "max_window_layers": 24,
122
+ "model_type": "qwen2",
123
+ "num_attention_heads": 14,
124
+ "num_hidden_layers": 24,
125
+ "num_key_value_heads": 2,
126
+ "rope_theta": 1000000.0,
127
+ "sliding_window": 32768,
128
+ "subnet_mode": "both",
129
+ "subnet_type": "None",
130
+ "temperature_attn": 1.3,
131
+ "temperature_mlp": 1.3,
132
+ "tie_word_embeddings": true,
133
+ "torch_dtype": "bfloat16",
134
+ "use_mrope": false,
135
+ "use_sliding_window": false,
136
+ "vocab_size": 151936
137
+ },
138
+ "threshold_connector": null,
139
+ "tokenizer_model_max_length": 2048,
140
+ "tokenizer_name_or_path": "Qwen/Qwen2.5-0.5B",
141
+ "tokenizer_padding_side": "right",
142
+ "tokenizer_use_fast": false,
143
+ "transformers_version": "4.40.1",
144
+ "tune_type_connector": "frozen",
145
+ "tune_type_llm": "frozen",
146
+ "tune_type_vision_tower": "frozen",
147
+ "tune_vision_tower_from_layer": -1,
148
+ "use_cache": false,
149
+ "vision_config": {
150
+ "hidden_act": "gelu_pytorch_tanh",
151
+ "hidden_size": 1152,
152
+ "image_size": 384,
153
+ "intermediate_size": 4304,
154
+ "layer_norm_eps": 1e-06,
155
+ "model_name_or_path": "google/siglip-so400m-patch14-384",
156
+ "model_name_or_path2": "",
157
+ "model_type": "siglip_vision_model",
158
+ "num_attention_heads": 16,
159
+ "num_hidden_layers": 27,
160
+ "patch_size": 14
161
+ },
162
+ "vision_feature_layer": -2,
163
+ "vision_feature_select_strategy": "patch",
164
+ "vision_hidden_size": 1152,
165
+ "vision_model_name_or_path": "google/siglip-so400m-patch14-384",
166
+ "vision_model_name_or_path2": "",
167
+ "vocab_size": 151936
168
+ }
169
+