| | import json |
| | import random |
| | from pathlib import Path |
| | from typing import Sequence, Callable |
| |
|
| | import fire |
| | import h5py |
| | import numpy as np |
| | from tqdm import tqdm |
| |
|
| | from infer_eval_utils import ( |
| | read_time_series_data, |
| | GENERATION_TASK_IDS, |
| | CLASSIFICATION_TASK_IDS, |
| | EVENT_DETECTION_TASK_IDS, |
| | ANOMALY_DETECTION_TASK_IDS, |
| | MCQ_TASK_IDS, |
| | IMPUTATION_TASK_IDS |
| | ) |
| |
|
| |
|
| | def read_raw_data(path: str | None) -> Sequence: |
| | if path is None: |
| | return [] |
| |
|
| | return read_time_series_data(path) |
| |
|
| |
|
| | def transform_raw_data_single_channel(raw_data: list | np.ndarray, |
| | dataset_id: str) -> str: |
| |
|
| | if dataset_id in GENERATION_TASK_IDS: |
| | data_str_list = [] |
| | for x in raw_data: |
| | if x != "X": |
| | data_str_list.append(f"{float(x):.3f}") |
| | else: |
| | data_str_list.append(x) |
| | data_text = " ".join(data_str_list) |
| | else: |
| | data_text = " ".join([f"{float(x):.3f}" for x in raw_data]) |
| |
|
| | return data_text |
| |
|
| |
|
| | def transform_raw_data_to_text(raw_data: list | np.ndarray, dataset_id: str, |
| | channel_detail: list[str]) -> str: |
| |
|
| | if isinstance(raw_data, np.ndarray): |
| | if raw_data.ndim > 1 and raw_data.shape[1] > 1 and len( |
| | channel_detail) == 0: |
| | channel_detail = [f"channel {i}" for i in range(raw_data.shape[1])] |
| |
|
| | if len(channel_detail) <= 1: |
| | data_text = transform_raw_data_single_channel(raw_data, dataset_id) |
| | else: |
| | data_text = "" |
| | for channel_idx, channel_name in enumerate(channel_detail): |
| | channel_data = raw_data[:, channel_idx] |
| | channel_data_text = transform_raw_data_single_channel( |
| | channel_data, dataset_id) |
| | data_text += f"{channel_name}: {channel_data_text} " |
| | else: |
| | data_text = transform_raw_data_single_channel(raw_data, dataset_id) |
| | return data_text |
| |
|
| |
|
| | def transform_gt_data_to_text(gt_data: list | np.ndarray, |
| | dataset_id: str) -> str: |
| | gt_data = np.array(gt_data) |
| | if gt_data.ndim == 1: |
| | data_text = transform_raw_data_single_channel(gt_data, dataset_id) |
| | else: |
| | data_text = "" |
| | for channel_idx in range(gt_data.shape[1]): |
| | channel_data = gt_data[:, channel_idx] |
| | channel_data_text = transform_raw_data_single_channel( |
| | channel_data, dataset_id) |
| | data_text += f"{channel_data_text}\n" |
| |
|
| | return data_text |
| |
|
| |
|
| | def get_extra_instruction(dataset_id: str, ) -> str: |
| | extra_instruction = "" |
| | if dataset_id == "ASU01_ASG02": |
| | extra_instruction = "Answer yes or no in the first line. If the Gravitational Wave is detected, answer the index of the starting time point in the second line." |
| | elif dataset_id == "EAU01_EAG02": |
| | extra_instruction = "Answer yes or no in the first line. If an Earthquake event is detected, answer the starting time point index of the P-wave in the second line, " \ |
| | "answer the starting time point index of the S-wave in the third line." |
| | elif dataset_id == "MFU01_MFU02": |
| | extra_instruction = "Output the diameter in the first line, and the position in the second line." |
| | elif dataset_id == "PHU01": |
| | extra_instruction = "Give each answer in a line. For example, if the answer is ['NORM', 'MI'], you should output: NORM\nMI." |
| | elif dataset_id == "MAG01": |
| | extra_instruction = "Give answer of each channel in a line so the number of predicted time points in each line should match the given one. For example, if " \ |
| | "it is required to predict the next 5 time points, and the predicted x0, x1, x2 are [[0.1, 0.2, 0.3, 0.4, 0.5], [0.4, 0.5, 0.6, 0.7, 0.8], [0.7, 0.8, 0.9, 0.1, 0.2]], " \ |
| | "you should output: 0.1 0.2 0.3 0.4 0.5\n0.4 0.5 0.6 0.7 0.8\n0.7 0.8 0.9 0.1 0.2." |
| | elif dataset_id in ANOMALY_DETECTION_TASK_IDS: |
| | extra_instruction = "Answer yes if anomaly points are detected, and no if there are only normal points." |
| | elif dataset_id in GENERATION_TASK_IDS: |
| | extra_instruction = "Output the values separated by spaces." |
| | return extra_instruction |
| |
|
| |
|
| | def extract_gt(data: dict, dataset_id: str) -> str | dict | Path: |
| | if dataset_id in CLASSIFICATION_TASK_IDS: |
| | gt = data["gt_result"]["gt_class"] |
| | if isinstance(gt, dict) and len(gt) == 1: |
| | gt = gt["default"] |
| | if isinstance(gt, list) and len(gt) == 1: |
| | gt = gt[0] |
| | elif dataset_id in GENERATION_TASK_IDS: |
| | gt = data["gt_ts"]["path"].strip("/") |
| | elif dataset_id in EVENT_DETECTION_TASK_IDS: |
| | gt = data["gt_result"] |
| | elif dataset_id in ANOMALY_DETECTION_TASK_IDS: |
| | gt = data["gt_result"]["contain"] |
| | elif dataset_id in MCQ_TASK_IDS: |
| | gt = data["gt_result"]["answer"] |
| | else: |
| | raise ValueError(f"Unsupported dataset id: {dataset_id}") |
| | return gt |
| |
|
| |
|
| | def initialize_model() -> Callable: |
| | """ |
| | Initialize the model here. The model can be called by: |
| | |
| | ```python |
| | response = model(prompt) |
| | # or |
| | response = model(prompt, max_tokens=max_tokens) # to limit the response length |
| | ``` |
| | """ |
| | pass |
| |
|
| |
|
| | def infer_dataset(model: Callable, dataset_data: list, scits_dir: Path, |
| | dataset_id: str, output_path: Path): |
| | print(f"Inferring {dataset_id}") |
| |
|
| | if dataset_id in GENERATION_TASK_IDS: |
| | ext = "h5" |
| | else: |
| | ext = "jsonl" |
| |
|
| | output_path = Path(output_path) / f"{dataset_id}.{ext}" |
| | output_path.parent.mkdir(parents=True, exist_ok=True) |
| | completed_ids = [] |
| |
|
| | if str(output_path).endswith(".jsonl"): |
| | has_metadata = False |
| |
|
| | if output_path.exists(): |
| | if str(output_path).endswith(".jsonl"): |
| | with open(output_path, 'r') as f: |
| | for line in f.readlines(): |
| | data = json.loads(line) |
| | if "id" in data: |
| | completed_ids.append(data["id"]) |
| | else: |
| | has_metadata = True |
| | elif str(output_path).endswith(".h5"): |
| | with h5py.File(output_path, 'r') as f: |
| | completed_ids = list(f.keys()) |
| |
|
| | completed_ids = set(completed_ids) |
| | random.shuffle(dataset_data) |
| | dataset_data = dataset_data[:10] |
| |
|
| | try: |
| | seq_length = dataset_data[0]["input_ts"]["length"] |
| | except: |
| | seq_length = None |
| |
|
| | for sample in tqdm(dataset_data): |
| | id = sample["id"].replace( |
| | "/", "%2F") |
| |
|
| | if id in completed_ids: |
| | continue |
| |
|
| | |
| | if sample["input_ts"] is None: |
| | raw_data_path = None |
| | channel_detail = None |
| | else: |
| | raw_data_path = scits_dir / sample["input_ts"]["path"].strip("/") |
| | channel_detail = sample["input_ts"]["channel_detail"] |
| |
|
| | raw_data = read_raw_data(raw_data_path) |
| | raw_data_text = transform_raw_data_to_text(raw_data, dataset_id, |
| | channel_detail) |
| |
|
| | gt = extract_gt(sample, dataset_id) |
| | extra_instruction = get_extra_instruction(dataset_id) |
| |
|
| | if dataset_id in GENERATION_TASK_IDS: |
| | |
| | gt_data = read_time_series_data(scits_dir / gt) |
| | gt_data_text = transform_gt_data_to_text( |
| | gt_data, dataset_id) |
| | max_tokens = len(gt_data_text) |
| | else: |
| | max_tokens = None |
| |
|
| | prompt_text = f'{sample["input_text"]} {extra_instruction} Give me the answer directly, ' \ |
| | f'without any other extra content (including punctuation). ' \ |
| | f'{raw_data_text}' |
| | output_text = model(text=prompt_text, max_tokens=max_tokens) |
| | |
| |
|
| | if dataset_id not in GENERATION_TASK_IDS: |
| | with open(output_path, 'a') as writer: |
| | if not has_metadata: |
| | metadata = {} |
| | if dataset_id in EVENT_DETECTION_TASK_IDS: |
| | metadata["seq_length"] = seq_length |
| | writer.write(json.dumps(metadata) + "\n") |
| | has_metadata = True |
| |
|
| | pred_results = output_text |
| | if "class_list" in sample["gt_result"] and isinstance( |
| | sample["gt_result"]["class_list"], dict) and len( |
| | sample["gt_result"]["class_list"]) > 1: |
| | if len(output_text.split("\n")) != len( |
| | sample["gt_result"]["class_list"]): |
| | pred_results = "NA" |
| | else: |
| | pred_results = {} |
| | for class_name, pred_result in zip( |
| | sample["gt_result"]["class_list"], |
| | output_text.split("\n")): |
| | pred_results[class_name] = pred_result |
| | writer.write( |
| | json.dumps({ |
| | "id": id, |
| | "output": pred_results, |
| | "ground_truth": gt |
| | }) + "\n") |
| | else: |
| | if "\n" not in output_text: |
| | pred_result = np.fromstring(output_text.strip(), |
| | dtype=np.float32, |
| | sep=' ') |
| | else: |
| | try: |
| | pred_result = np.vstack([ |
| | np.fromstring(x.strip(), dtype=np.float32, sep=' ') |
| | for x in output_text.split("\n") |
| | ]).transpose() |
| | except ValueError: |
| | pred_result = np.array([]) |
| |
|
| | with h5py.File(output_path, 'a') as writer: |
| | writer[f"{id}/pred_result"] = pred_result |
| | writer[f"{id}/gt_path"] = gt.__str__().encode("utf-8") |
| | if dataset_id in IMPUTATION_TASK_IDS: |
| | writer[f"{id}/input_ts_path"] = sample["input_ts"][ |
| | "path"].strip("/").encode("utf-8") |
| | if "base_path" not in writer: |
| | writer["base_path"] = scits_dir.__str__().encode("utf-8") |
| |
|
| |
|
| |
|
| | def infer( |
| | scits_dir: str, |
| | output_dir: str, |
| | ): |
| |
|
| | |
| | model: Callable = initialize_model() |
| |
|
| | scits_dir = Path(scits_dir) |
| | output_dir = Path(output_dir) |
| | dataset_data = [] |
| | prev_dataset_id = None |
| | with open(scits_dir / "meta_data.jsonl", 'r') as f: |
| | for line in f.readlines(): |
| | sample = json.loads(line) |
| | dataset_id = "_".join(sample["task_id"]) |
| |
|
| | if dataset_id != prev_dataset_id: |
| | if prev_dataset_id is not None: |
| | infer_dataset(model, dataset_data, scits_dir, prev_dataset_id, |
| | output_dir) |
| | dataset_data = [] |
| | prev_dataset_id = dataset_id |
| | dataset_data.append(sample) |
| |
|
| | infer_dataset(model, dataset_data, scits_dir, prev_dataset_id, output_dir) |
| |
|
| |
|
| | if __name__ == '__main__': |
| | fire.Fire(infer) |
| |
|