repo_id
stringlengths
6
101
size
int64
367
5.14M
file_path
stringlengths
2
269
content
stringlengths
367
5.14M
2977094657/DidaAPI
7,837
routers/tasks.py
"""任务相关API路由""" from fastapi import APIRouter, HTTPException, Query, Body from typing import Optional from models import ApiResponse from services import dida_service from utils import app_logger router = APIRouter(prefix="/tasks", tags=["任务管理"]) @router.post("/set-auth", response_model=ApiResponse, summary="设置认证会话", description="设置滴答清单API的认证令牌,用于后续API调用") async def set_auth_session( auth_token: str = Body(..., description="认证令牌(t cookie值)"), csrf_token: str = Body(..., description="CSRF令牌") ) -> ApiResponse: """ 设置认证会话 - **auth_token**: 认证令牌(从微信登录获得的 t cookie值) - **csrf_token**: CSRF令牌(从微信登录获得的 _csrf_token值) 设置后可以调用其他需要认证的API接口 """ try: app_logger.info("设置认证会话") if not auth_token or not csrf_token: raise HTTPException( status_code=400, detail="认证令牌和CSRF令牌不能为空" ) session_id = dida_service.set_auth_session(auth_token, csrf_token) return ApiResponse( code=200, message="认证会话设置成功", data={ "session_id": session_id, "status": "已设置认证会话,可以调用其他API" } ) except HTTPException: raise except Exception as e: app_logger.error(f"设置认证会话时发生错误: {e}") raise HTTPException( status_code=500, detail=f"设置认证会话失败: {str(e)}" ) @router.get("/all", summary="获取所有任务", description="获取当前用户的所有任务列表") async def get_all_tasks(): """ 获取所有任务 返回当前用户的所有任务列表,包括: - 任务ID、标题、内容 - 任务状态(0=未完成,2=已完成) - 优先级、创建时间、修改时间 - 项目ID、标签等信息 **注意**: 需要先调用 `/tasks/set-auth` 设置认证会话 """ try: app_logger.info("请求获取所有任务") result = await dida_service.get_all_tasks() if not result: return {"error": "获取任务失败,请稍后重试"} # 记录日志 if 'error' in result: app_logger.info(f"获取任务失败: {result.get('error')}") else: app_logger.info(f"任务获取完成") # 直接返回原始响应 return result except HTTPException: raise except Exception as e: app_logger.error(f"获取任务时发生未知错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/summary", response_model=ApiResponse, summary="获取任务统计", description="获取任务的统计信息") async def get_tasks_summary() -> ApiResponse: """ 获取任务统计 返回任务的统计信息: - 总任务数 - 已完成任务数 - 未完成任务数 - 完成率等 """ try: app_logger.info("请求获取任务统计") result = await dida_service.get_all_tasks() if not result or "error" in result: return {"error": "获取任务统计失败", "details": result} # 统计任务信息 - 从原始响应中提取 total_tasks = 0 completed_tasks = 0 pending_tasks = 0 # 解析原始响应数据 if isinstance(result, dict) and 'syncTaskBean' in result: task_data = result['syncTaskBean'] if 'update' in task_data: raw_tasks = task_data['update'] total_tasks = len(raw_tasks) for task in raw_tasks: if task.get('status') == 2: # 已完成 completed_tasks += 1 else: # 未完成 pending_tasks += 1 completion_rate = (completed_tasks / total_tasks * 100) if total_tasks > 0 else 0 return ApiResponse( code=200, message="获取任务统计成功", data={ "total_tasks": total_tasks, "completed_tasks": completed_tasks, "pending_tasks": pending_tasks, "completion_rate": round(completion_rate, 2) } ) except HTTPException: raise except Exception as e: app_logger.error(f"获取任务统计时发生错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/completed", summary="获取已完成/已放弃任务", description="获取已完成或已放弃的任务列表,支持分页获取") async def get_completed_tasks( to: Optional[str] = Query(None, description="分页参数,使用上次响应最后一个任务的completedTime字段,格式:2025-03-15T13:30:54.000+0000"), status: str = Query("Completed", description="任务状态:Completed(已完成) 或 Abandoned(已放弃)") ): """ 获取已完成/已放弃任务 支持分页获取已完成或已放弃的任务列表: - **第一次请求**: 不传to参数,获取最新的任务 - **后续请求**: 使用上次响应最后一个任务的completedTime字段作为to参数 - **状态选择**: 通过status参数选择获取已完成(Completed)或已放弃(Abandoned)的任务 **分页机制说明**: 1. completedTime原始格式:2025-03-15T13:30:54.000+0000 2. to参数传入:使用completedTime的原始格式 3. API内部转换:2025-03-15T13:30:54.000+0000 → 2025-03-15 13:30:54 4. URL示例:https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Completed **状态参数说明**: - Completed: 获取已完成的任务 - Abandoned: 获取已放弃的任务 **注意**: 需要先完成微信登录获取认证会话 """ try: app_logger.info(f"请求获取任务,状态: {status},分页参数: {to}") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"} # 调用服务获取任务 result = await dida_service.get_completed_tasks(to, status) if not result: return {"error": "service_error", "message": f"获取{status}任务失败,请稍后重试"} # 记录日志 if 'error' in result: app_logger.info(f"{status}任务获取失败: {result.get('error')}") else: task_count = len(result) if isinstance(result, list) else 0 app_logger.info(f"{status}任务获取完成,任务数: {task_count}") # 如果有任务,记录最后一个任务的completedTime,便于下次分页 if isinstance(result, list) and len(result) > 0: last_task = result[-1] last_completed_time = last_task.get('completedTime') if last_completed_time: app_logger.info(f"最后一个任务的completedTime: {last_completed_time}") # 直接返回原始响应 return result except Exception as e: app_logger.error(f"获取{status}任务时发生未知错误: {e}") return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"} @router.get("/trash", summary="获取垃圾桶任务", description="获取垃圾桶中的任务列表") async def get_trash_tasks( limit: int = Query(50, description="每页任务数量,默认50"), task_type: int = Query(1, description="任务类型,默认1") ): """ 获取垃圾桶任务 获取垃圾桶中的任务列表: - **limit**: 每页返回的任务数量,默认50 - **task_type**: 任务类型,默认1 **响应格式**: ```json { "tasks": [...], // 任务列表 "next": 0 // 下一页标识 } ``` **注意**: 需要先完成微信登录获取认证会话 """ try: app_logger.info(f"请求获取垃圾桶任务,limit: {limit}, type: {task_type}") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"} # 调用服务获取垃圾桶任务 result = await dida_service.get_trash_tasks(limit, task_type) if not result: return {"error": "service_error", "message": "获取垃圾桶任务失败,请稍后重试"} # 记录日志 if 'error' in result: app_logger.info(f"垃圾桶任务获取失败: {result.get('error')}") else: task_count = len(result.get('tasks', [])) if isinstance(result, dict) else 0 next_page = result.get('next', 0) if isinstance(result, dict) else 0 app_logger.info(f"垃圾桶任务获取完成,任务数: {task_count}, next: {next_page}") # 直接返回原始响应 return result except Exception as e: app_logger.error(f"获取垃圾桶任务时发生未知错误: {e}") return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
294coder/Efficient-MIF
1,623
configs/hyper_transformer_config.yaml
optimizer: name: 'adam' lr: !!float 1e-5 weight_decay: !!float 1e-6 lr_scheduler: # name: 'multi_step' # milestones: [] # gamma: 0.1 name: 'identity' max_norm: null path: # train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5' # val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5' train_path: '/Data3/cao/cavex4/test_cavepatches128-2.h5' # train_path: "/media/office-401/Elements SE/cao/cavex4/test_cavepatches64-2.h5" # train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5" val_path: "/Data3/cao/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5" # train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/Chikusei_x4/train_Chikusei.h5" # train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/Chikusei_x4/test_chikuseipatches340.h5" # val_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/Chikusei_x4/test_Chikusei.h5" # train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Pavia/Train_Pavia.h5" # val_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Pavia/Test_Pavia.h5" # train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Botswana/Train_Botswana.h5" # val_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Botswana/Test_Botswana.h5" network_configs: config: none logger_config: base_path: ./log_file/ name: hyper_transformer file_mode: w
294coder/Efficient-MIF
1,412
configs/hpmnet_config.yaml
optimizer: name: "adamw" lr: !!float 1e-4 weight_decay: !!float 1e-6 lr_scheduler: name: "multi_step" milestones: [1000, 1600] gamma: 1 max_norm: 0.01 path: # train_path: "/Data2/ZiHanCao/datasets/pansharpening/GF5_GF1/train_GF5_GF1.h5" # val_path: "/Data2/Datasets/GF5_GF1/test_GF5_GF1-new.h5" # train_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5" # val_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5" # train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5" # val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5" # train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5" # train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5" # val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5" train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5" val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5" network_configs: n_iter: 6 h_nc: 64 in_c: 9 out_c: 8 m_c: 3 nc: [80, 160, 320] nb: 1 act_mode: "R" downsample_mode: "strideconv" upsample_mode: "convtranspose" logger_config: base_path: ./log_file/ name: hpmnet_gf2 file_mode: w
294coder/Efficient-MIF
2,272
configs/MGDN_config.yaml
optimizer: name: "adamw" lr: !!float 1e-3 weight_decay: !!float 1e-6 lr_scheduler: name: 'cos_anneal_restart_reduce' T_0: 70 T_mult: 2 lr_mult: 0.9 eta_min: !!float 8e-5 max_norm: 0.3 ema_decay: 0.99 path: wv3_train_path: "/Data3/cao/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5" wv3_val_path: "/Data3/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5" gf2_train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5" gf2_val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5" qb_train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5" qb_val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5" cave_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/train_cave(with_up)x4.h5" cave_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5" cave_x8_train_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/train_cave(with_up)x8_rgb_16.h5" cave_x8_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/test_cave(with_up)x8_rgb.h5" harvard_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5" harvard_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/test_harvard(with_up)x4_rgb.h5" harvard_x8_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/train_harvard(with_up)x8_rgb.h5" harvard_x8_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/test_harvard(with_up)x8_rgb.h5" msrs_base_dir: "/Data3/cao/ZiHanCao/datasets/MSRS" llvip_base_dir: "/Data3/cao/ZiHanCao/datasets/LLVIP/data" # the mask is generated by using Ground-DINO-SAM project tno_base_dir: "/Data3/cao/ZiHanCao/datasets/TNO" roadscene_base_dir: "/Data3/cao/ZiHanCao/datasets/RoadSceneFusion" roadscene_tno_joint_base_dir: "/Data3/cao/ZiHanCao/datasets/RoadScene_and_TNO" med_harvard_base_dir: "/Data3/cao/ZiHanCao/datasets/MedHarvard" network_configs: GF_chans: 60 embed_dim: 60 depths: - 6 - 6 num_heads: - 6 - 6 window_size: 8 kernel_num: - 1 - 1 - 1 mlp_ratio: 2.0 inference: false logger_config: base_path: ./log_file/ name: MGDN file_mode: w
2977094657/DidaAPI
9,036
routers/export.py
"""自定义导出功能API路由""" from fastapi import APIRouter, HTTPException from fastapi.responses import StreamingResponse import io import urllib.parse from services.export_service import export_service from services.dida_service import dida_service from utils import app_logger router = APIRouter(prefix="/custom", tags=["自定义接口"]) @router.get("/export/tasks/excel", summary="导出任务到Excel", description="导出所有任务到Excel文件,包含全部任务、已完成任务、放弃任务、垃圾桶任务四个工作表") async def export_tasks_to_excel(): """ 导出任务到Excel 将用户的所有任务导出为Excel文件,包含以下工作表: - **全部任务**: 当前所有任务(未完成和已完成) - **已完成任务**: 历史已完成的任务 - **放弃任务**: 历史放弃的任务 - **垃圾桶任务**: 已删除的任务 每个工作表包含任务的完整字段信息,包括: - 基本信息:任务ID、标题、内容、描述、项目信息 - 状态信息:任务状态、优先级、完成进度 - 时间信息:创建时间、修改时间、开始日期、截止日期 - 重复设置:重复标志、重复来源、首次重复日期 - 提醒设置:提醒配置、排除日期 - 层级关系:父任务、子任务关系 - 其他属性:标签、附件、评论数量等 **注意**: 需要先调用认证接口设置会话 """ try: app_logger.info("请求导出任务到Excel") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: raise HTTPException( status_code=401, detail="未设置认证会话,请先完成登录" ) # 调用导出服务 result = await export_service.export_tasks_to_excel() if 'error' in result: app_logger.error(f"导出任务失败: {result['error']}") raise HTTPException( status_code=500, detail=f"导出失败: {result['error']}" ) app_logger.info(f"任务导出成功,文件大小: {result['size']} 字节") # 对文件名进行URL编码以支持中文 encoded_filename = urllib.parse.quote(result['filename'], safe='') # 返回文件下载响应 return StreamingResponse( io.BytesIO(result['content']), media_type=result['content_type'], headers={ "Content-Disposition": f"attachment; filename*=UTF-8''{encoded_filename}" } ) except HTTPException: raise except Exception as e: app_logger.error(f"导出任务时发生未知错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/export/tasks/excel/info", summary="获取任务导出信息", description="获取当前用户任务的统计信息,用于导出前预览") async def get_export_info(): """ 获取任务导出信息 返回当前用户任务的统计信息,包括: - 全部任务数量 - 已完成任务数量 - 垃圾桶任务数量 - 认证状态 **注意**: 需要先调用认证接口设置会话 """ try: app_logger.info("请求获取任务导出信息") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return { "error": "no_auth_session", "message": "未设置认证会话,请先完成登录", "auth_status": False } # 获取各类任务统计 stats = { "auth_status": True, "all_tasks_count": 0, "completed_tasks_count": 0, "abandoned_tasks_count": 0, "trash_tasks_count": 0, "session_info": session_status } # 获取全部任务统计 try: all_tasks_result = await dida_service.get_all_tasks() if all_tasks_result and 'error' not in all_tasks_result: tasks = all_tasks_result.get('syncTaskBean', {}).get('update', []) stats["all_tasks_count"] = len(tasks) except Exception as e: app_logger.warning(f"获取全部任务统计失败: {e}") # 获取已完成任务统计(只获取第一页用于统计) try: completed_result = await dida_service.get_completed_tasks(None, "Completed") if completed_result and 'error' not in completed_result: if isinstance(completed_result, list): stats["completed_tasks_count"] = len(completed_result) except Exception as e: app_logger.warning(f"获取已完成任务统计失败: {e}") # 获取放弃任务统计 try: abandoned_tasks_result = await dida_service.get_completed_tasks(None, "Abandoned") if abandoned_tasks_result and 'error' not in abandoned_tasks_result: if isinstance(abandoned_tasks_result, list): stats["abandoned_tasks_count"] = len(abandoned_tasks_result) except Exception as e: app_logger.warning(f"获取放弃任务统计失败: {e}") # 获取垃圾桶任务统计 try: trash_result = await dida_service.get_trash_tasks() if trash_result and 'error' not in trash_result: tasks = trash_result.get('tasks', []) stats["trash_tasks_count"] = len(tasks) except Exception as e: app_logger.warning(f"获取垃圾桶任务统计失败: {e}") app_logger.info(f"任务统计获取完成: {stats}") return stats except Exception as e: app_logger.error(f"获取任务导出信息时发生错误: {e}") return { "error": "server_error", "message": f"服务器内部错误: {str(e)}", "auth_status": False } @router.get("/export/focus/excel", summary="导出专注记录到Excel", description="导出所有专注记录到Excel文件,包含完整的专注时间线数据") async def export_focus_records_to_excel(): """ 导出专注记录到Excel 将用户的所有专注记录导出为Excel文件,包含: - **专注记录时间线**: 所有专注记录的详细信息 每个工作表包含专注记录的完整字段信息,包括: - 基本信息:专注记录ID、开始时间、结束时间、创建时间 - 专注状态:专注状态、专注时长、暂停时长、实际专注时长 - 任务信息:任务ID、任务标题、项目ID、项目名称 - 标签信息:标签列表、标签ID列表 - 设备信息:设备类型、平台、应用版本 - 专注模式:专注模式、番茄钟时长、休息时长 - 其他属性:用户ID、时区、删除状态等 **注意**: - 需要先调用认证接口设置会话 - 会自动分页获取所有历史专注记录 """ try: app_logger.info("请求导出专注记录到Excel") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: raise HTTPException( status_code=401, detail="未设置认证会话,请先完成登录" ) # 调用导出服务 result = await export_service.export_focus_records_to_excel() if 'error' in result: app_logger.error(f"导出专注记录失败: {result['error']}") raise HTTPException( status_code=500, detail=f"导出失败: {result['error']}" ) app_logger.info(f"专注记录导出成功,文件大小: {result['size']} 字节") # 对文件名进行URL编码以支持中文 encoded_filename = urllib.parse.quote(result['filename'], safe='') # 返回文件下载响应 return StreamingResponse( io.BytesIO(result['content']), media_type=result['content_type'], headers={ "Content-Disposition": f"attachment; filename*=UTF-8''{encoded_filename}" } ) except HTTPException: raise except Exception as e: app_logger.error(f"导出专注记录时发生未知错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/export/focus/excel/info", summary="获取专注记录导出信息", description="获取当前用户专注记录的统计信息,用于导出前预览") async def get_focus_export_info(): """ 获取专注记录导出信息 返回当前用户专注记录的统计信息,包括: - 专注记录总数量(预估) - 认证状态 **注意**: 需要先调用认证接口设置会话 """ try: app_logger.info("请求获取专注记录导出信息") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return { "error": "no_auth_session", "message": "未设置认证会话,请先完成登录", "auth_status": False } # 获取专注记录统计(只获取第一页用于预估) stats = { "auth_status": True, "focus_records_count_estimate": 0, "session_info": session_status, "note": "专注记录数量为预估值,实际导出时会获取所有历史数据" } try: # 获取认证信息 current_session = dida_service.current_session auth_token = current_session['auth_token'] csrf_token = current_session['csrf_token'] # 获取第一页专注记录用于预估 from services.pomodoro_service import pomodoro_service result = await pomodoro_service.get_focus_timeline(auth_token, csrf_token, None) if result and 'error' not in result: if isinstance(result, list): # 如果第一页有31条记录,预估可能有更多数据 first_page_count = len(result) if first_page_count >= 31: stats["focus_records_count_estimate"] = f"{first_page_count}+ (需要分页获取完整数据)" else: stats["focus_records_count_estimate"] = first_page_count else: stats["focus_records_count_estimate"] = "无法预估" except Exception as e: app_logger.warning(f"获取专注记录统计失败: {e}") stats["focus_records_count_estimate"] = "获取失败" app_logger.info(f"专注记录统计获取完成: {stats}") return stats except Exception as e: app_logger.error(f"获取专注记录导出信息时发生错误: {e}") return { "error": "server_error", "message": f"服务器内部错误: {str(e)}", "auth_status": False }
2977094657/DidaAPI
5,669
routers/system.py
"""系统相关API路由""" from fastapi import APIRouter from typing import Dict, Any from core import urls from models import ApiResponse from utils import app_logger router = APIRouter(prefix="/system", tags=["系统管理"]) @router.get("/urls", response_model=ApiResponse, summary="获取所有URL配置", description="获取项目中管理的所有URL和外部链接配置") async def get_all_urls() -> ApiResponse: """ 获取所有URL配置 返回项目中统一管理的所有URL和外部链接 """ try: all_urls = urls.get_all_external_urls() api_endpoints = urls.get_api_endpoints() return ApiResponse( code=200, message="获取URL配置成功", data={ "external_urls": all_urls, "api_endpoints": api_endpoints } ) except Exception as e: app_logger.error(f"获取URL配置时发生错误: {e}") return ApiResponse( code=500, message=f"获取URL配置失败: {str(e)}", data=None ) @router.get("/urls/wechat", response_model=ApiResponse, summary="获取微信相关URL", description="获取微信开放平台相关的所有URL配置") async def get_wechat_urls() -> ApiResponse: """ 获取微信相关URL 返回微信开放平台相关的URL配置 """ try: return ApiResponse( code=200, message="获取微信URL配置成功", data={ "wechat_urls": urls.WECHAT_URLS, "wechat_config": urls.WECHAT_CONFIG, "helper_functions": [ "build_wechat_qr_url(state)", "build_wechat_poll_url(uuid, timestamp)", "build_wechat_validate_url(code, state)" ] } ) except Exception as e: app_logger.error(f"获取微信URL配置时发生错误: {e}") return ApiResponse( code=500, message=f"获取微信URL配置失败: {str(e)}", data=None ) @router.get("/urls/dida", response_model=ApiResponse, summary="获取滴答清单API URL", description="获取滴答清单API相关的所有URL配置") async def get_dida_urls() -> ApiResponse: """ 获取滴答清单API URL 返回滴答清单API相关的URL配置 """ try: return ApiResponse( code=200, message="获取滴答清单URL配置成功", data={ "base_config": urls.DIDA_API_BASE, "auth_apis": urls.DIDA_AUTH_APIS, "task_apis": urls.DIDA_TASK_APIS, "project_apis": urls.DIDA_PROJECT_APIS, "statistics_apis": urls.DIDA_STATISTICS_APIS, "pomodoro_apis": urls.DIDA_POMODORO_APIS, "habit_apis": urls.DIDA_HABIT_APIS, "helper_functions": [ "build_dida_api_url(endpoint)" ] } ) except Exception as e: app_logger.error(f"获取滴答清单URL配置时发生错误: {e}") return ApiResponse( code=500, message=f"获取滴答清单URL配置失败: {str(e)}", data=None ) @router.get("/urls/docs", response_model=ApiResponse, summary="获取文档链接", description="获取官方文档和技术参考链接") async def get_doc_urls() -> ApiResponse: """ 获取文档链接 返回官方文档和技术参考链接 """ try: return ApiResponse( code=200, message="获取文档链接成功", data={ "official_docs": urls.OFFICIAL_DOCS, "tech_references": urls.TECH_REFERENCES, "project_docs": { "frontend_docs": "http://localhost:3000", "api_docs": "http://localhost:3000/api/", "url_management": "http://localhost:3000/api/url-management", "wechat_flow": "http://localhost:3000/api/auth/wechat-login-flow" } } ) except Exception as e: app_logger.error(f"获取文档链接时发生错误: {e}") return ApiResponse( code=500, message=f"获取文档链接失败: {str(e)}", data=None ) @router.get("/info", response_model=ApiResponse, summary="获取系统信息", description="获取项目的基本信息和配置概览") async def get_system_info() -> ApiResponse: """ 获取系统信息 返回项目的基本信息和配置概览 """ try: from core import config return ApiResponse( code=200, message="获取系统信息成功", data={ "project_name": "滴答清单Web端API接口", "version": "1.0.0", "description": "滴答清单原始API的封装层,提供简化的接口和完整的API文档", "features": [ "微信扫码登录", "任务管理", "会话持久化", "统一URL管理", "完整API文档" ], "url_management": { "total_external_urls": len(urls.WECHAT_URLS) + len(urls.DIDA_API_BASE) + len(urls.OFFICIAL_DOCS) + len(urls.TECH_REFERENCES), "total_api_endpoints": len(urls.DIDA_AUTH_APIS) + len(urls.DIDA_TASK_APIS) + len(urls.DIDA_PROJECT_APIS) + len(urls.DIDA_STATISTICS_APIS) + len(urls.DIDA_POMODORO_APIS) + len(urls.DIDA_HABIT_APIS), "management_file": "core/urls.py" }, "config": { "app": config.app, "request_config": config.get('request_config', {}), "database": config.database } } ) except Exception as e: app_logger.error(f"获取系统信息时发生错误: {e}") return ApiResponse( code=500, message=f"获取系统信息失败: {str(e)}", data=None )
281677160/openwrt-package
66,136
luci-app-passwall/root/usr/share/passwall/subscribe.lua
#!/usr/bin/lua ------------------------------------------------ -- @author William Chan <root@williamchan.me> ------------------------------------------------ require 'luci.util' require 'luci.jsonc' require 'luci.sys' local appname = 'passwall' local api = require ("luci.passwall.api") local datatypes = require "luci.cbi.datatypes" -- these global functions are accessed all the time by the event handler -- so caching them is worth the effort local tinsert = table.insert local ssub, slen, schar, sbyte, sformat, sgsub = string.sub, string.len, string.char, string.byte, string.format, string.gsub local split = api.split local jsonParse, jsonStringify = luci.jsonc.parse, luci.jsonc.stringify local base64Decode = api.base64Decode local uci = api.uci local fs = api.fs uci:revert(appname) local has_ss = api.is_finded("ss-redir") local has_ss_rust = api.is_finded("sslocal") local has_ssr = api.is_finded("ssr-local") and api.is_finded("ssr-redir") local has_trojan_plus = api.is_finded("trojan-plus") local has_singbox = api.finded_com("sing-box") local has_xray = api.finded_com("xray") local has_hysteria2 = api.finded_com("hysteria") local allowInsecure_default = nil -- 取节点使用core类型(节点订阅页面未设置时,自动取默认) local function get_core(field, candidates) local v = uci:get(appname, "@global_subscribe[0]", field) if not v or v == "" then for _, c in ipairs(candidates) do if c[1] then return c[2] end end end return v end local ss_type_default = get_core("ss_type", {{has_ss,"shadowsocks-libev"},{has_ss_rust,"shadowsocks-rust"},{has_singbox,"sing-box"},{has_xray,"xray"}}) local trojan_type_default = get_core("trojan_type", {{has_trojan_plus,"trojan-plus"},{has_singbox,"sing-box"},{has_xray,"xray"}}) local vmess_type_default = get_core("vmess_type", {{has_xray,"xray"},{has_singbox,"sing-box"}}) local vless_type_default = get_core("vless_type", {{has_xray,"xray"},{has_singbox,"sing-box"}}) local hysteria2_type_default = get_core("hysteria2_type", {{has_hysteria2,"hysteria2"},{has_singbox,"sing-box"}}) ---- local domain_strategy_default = uci:get(appname, "@global_subscribe[0]", "domain_strategy") or "" local domain_strategy_node = "" local preproxy_node_group, to_node_group, chain_node_type = "", "", "" -- 判断是否过滤节点关键字 local filter_keyword_mode_default = uci:get(appname, "@global_subscribe[0]", "filter_keyword_mode") or "0" local filter_keyword_discard_list_default = uci:get(appname, "@global_subscribe[0]", "filter_discard_list") or {} local filter_keyword_keep_list_default = uci:get(appname, "@global_subscribe[0]", "filter_keep_list") or {} local function is_filter_keyword(value) if filter_keyword_mode_default == "1" then for k,v in ipairs(filter_keyword_discard_list_default) do if value:find(v, 1, true) then return true end end elseif filter_keyword_mode_default == "2" then local result = true for k,v in ipairs(filter_keyword_keep_list_default) do if value:find(v, 1, true) then result = false end end return result elseif filter_keyword_mode_default == "3" then local result = false for k,v in ipairs(filter_keyword_discard_list_default) do if value:find(v, 1, true) then result = true end end for k,v in ipairs(filter_keyword_keep_list_default) do if value:find(v, 1, true) then result = false end end return result elseif filter_keyword_mode_default == "4" then local result = true for k,v in ipairs(filter_keyword_keep_list_default) do if value:find(v, 1, true) then result = false end end for k,v in ipairs(filter_keyword_discard_list_default) do if value:find(v, 1, true) then result = true end end return result end return false end local nodeResult = {} -- update result local isDebug = false local log = function(...) if isDebug == true then local result = os.date("%Y-%m-%d %H:%M:%S: ") .. table.concat({...}, " ") print(result) else api.log(...) end end local nodes_table = {} for k, e in ipairs(api.get_valid_nodes()) do if e.node_type == "normal" then nodes_table[#nodes_table + 1] = e end end -- 获取各项动态配置的当前服务器,可以用 get 和 set, get必须要获取到节点表 local CONFIG = {} do local function import_config(protocol) local name = string.upper(protocol) local szType = "@global[0]" local option = protocol .. "_node" local node_id = uci:get(appname, szType, option) CONFIG[#CONFIG + 1] = { log = true, remarks = name .. "节点", currentNode = node_id and uci:get_all(appname, node_id) or nil, set = function(o, server) uci:set(appname, szType, option, server) o.newNodeId = server end } end import_config("tcp") import_config("udp") if true then local i = 0 local option = "node" uci:foreach(appname, "socks", function(t) i = i + 1 local id = t[".name"] local node_id = t[option] CONFIG[#CONFIG + 1] = { log = true, id = id, remarks = "Socks节点列表[" .. i .. "]", currentNode = node_id and uci:get_all(appname, node_id) or nil, set = function(o, server) if not server or server == "" then if #nodes_table > 0 then server = nodes_table[1][".name"] end end uci:set(appname, t[".name"], option, server) o.newNodeId = server end } if t.autoswitch_backup_node and #t.autoswitch_backup_node > 0 then local flag = "Socks节点列表[" .. i .. "]备用节点的列表" local currentNodes = {} local newNodes = {} for k, node_id in ipairs(t.autoswitch_backup_node) do if node_id then local currentNode = uci:get_all(appname, node_id) or nil if currentNode then currentNodes[#currentNodes + 1] = { log = true, remarks = flag .. "[" .. k .. "]", currentNode = currentNode, set = function(o, server) if server and server ~= "nil" then table.insert(o.newNodes, server) end end } end end end CONFIG[#CONFIG + 1] = { remarks = flag, currentNodes = currentNodes, newNodes = newNodes, set = function(o, newNodes) if o then if not newNodes then newNodes = o.newNodes end uci:set_list(appname, id, "autoswitch_backup_node", newNodes or {}) end end } end end) end if true then local i = 0 local option = "lbss" local function is_ip_port(str) if type(str) ~= "string" then return false end local ip, port = str:match("^([%d%.]+):(%d+)$") return ip and datatypes.ipaddr(ip) and tonumber(port) and tonumber(port) <= 65535 end uci:foreach(appname, "haproxy_config", function(t) i = i + 1 local node_id = t[option] CONFIG[#CONFIG + 1] = { log = true, id = t[".name"], remarks = "HAProxy负载均衡节点列表[" .. i .. "]", currentNode = node_id and uci:get_all(appname, node_id) or nil, set = function(o, server) -- 如果当前 lbss 值不是 ip:port 格式,才进行修改 if not is_ip_port(t[option]) then uci:set(appname, t[".name"], option, server) o.newNodeId = server end end, delete = function(o) -- 如果当前 lbss 值不是 ip:port 格式,才进行删除 if not is_ip_port(t[option]) then uci:delete(appname, t[".name"]) end end } end) end if true then local i = 0 local options = {"tcp", "udp"} uci:foreach(appname, "acl_rule", function(t) i = i + 1 for index, value in ipairs(options) do local option = value .. "_node" local node_id = t[option] CONFIG[#CONFIG + 1] = { log = true, id = t[".name"], remarks = "访问控制列表[" .. i .. "]", currentNode = node_id and uci:get_all(appname, node_id) or nil, set = function(o, server) uci:set(appname, t[".name"], option, server) o.newNodeId = server end } end end) end uci:foreach(appname, "nodes", function(node) local node_id = node[".name"] if node.protocol and node.protocol == '_shunt' then local rules = {} uci:foreach(appname, "shunt_rules", function(e) if e[".name"] and e.remarks then table.insert(rules, e) end end) table.insert(rules, { [".name"] = "default_node", remarks = "默认" }) table.insert(rules, { [".name"] = "main_node", remarks = "默认前置" }) for k, e in pairs(rules) do local _node_id = node[e[".name"]] or nil if _node_id and api.parseURL(_node_id) then else CONFIG[#CONFIG + 1] = { log = false, currentNode = _node_id and uci:get_all(appname, _node_id) or nil, remarks = "分流" .. e.remarks .. "节点", set = function(o, server) if not server then server = "" end uci:set(appname, node_id, e[".name"], server) o.newNodeId = server end } end end elseif node.protocol and node.protocol == '_balancing' then local flag = "Xray负载均衡节点[" .. node_id .. "]列表" local currentNodes = {} local newNodes = {} if node.balancing_node then for k, node in pairs(node.balancing_node) do currentNodes[#currentNodes + 1] = { log = true, node = node, currentNode = node and uci:get_all(appname, node) or nil, remarks = node, set = function(o, server) if o and server and server ~= "nil" then table.insert(o.newNodes, server) end end } end end CONFIG[#CONFIG + 1] = { remarks = flag, currentNodes = currentNodes, newNodes = newNodes, set = function(o, newNodes) if o then if not newNodes then newNodes = o.newNodes end uci:set_list(appname, node_id, "balancing_node", newNodes or {}) end end } --后备节点 local currentNode = uci:get_all(appname, node_id) or nil if currentNode and currentNode.fallback_node then CONFIG[#CONFIG + 1] = { log = true, id = node_id, remarks = "Xray负载均衡节点[" .. node_id .. "]后备节点", currentNode = uci:get_all(appname, currentNode.fallback_node) or nil, set = function(o, server) uci:set(appname, node_id, "fallback_node", server) o.newNodeId = server end, delete = function(o) uci:delete(appname, node_id, "fallback_node") end } end elseif node.protocol and node.protocol == '_urltest' then local flag = "Sing-Box URLTest节点[" .. node_id .. "]列表" local currentNodes = {} local newNodes = {} if node.urltest_node then for k, node in pairs(node.urltest_node) do currentNodes[#currentNodes + 1] = { log = true, node = node, currentNode = node and uci:get_all(appname, node) or nil, remarks = node, set = function(o, server) if o and server and server ~= "nil" then table.insert(o.newNodes, server) end end } end end CONFIG[#CONFIG + 1] = { remarks = flag, currentNodes = currentNodes, newNodes = newNodes, set = function(o, newNodes) if o then if not newNodes then newNodes = o.newNodes end uci:set_list(appname, node_id, "urltest_node", newNodes or {}) end end } else --前置代理节点 local currentNode = uci:get_all(appname, node_id) or nil if currentNode and currentNode.preproxy_node then CONFIG[#CONFIG + 1] = { log = true, id = node_id, remarks = "节点[" .. node_id .. "]前置代理节点", currentNode = uci:get_all(appname, currentNode.preproxy_node) or nil, set = function(o, server) uci:set(appname, node_id, "preproxy_node", server) o.newNodeId = server end, delete = function(o) uci:delete(appname, node_id, "preproxy_node") end } end --落地节点 local currentNode = uci:get_all(appname, node_id) or nil if currentNode and currentNode.to_node then CONFIG[#CONFIG + 1] = { log = true, id = node_id, remarks = "节点[" .. node_id .. "]落地节点", currentNode = uci:get_all(appname, currentNode.to_node) or nil, set = function(o, server) uci:set(appname, node_id, "to_node", server) o.newNodeId = server end, delete = function(o) uci:delete(appname, node_id, "to_node") end } end end end) for k, v in pairs(CONFIG) do if v.currentNodes and type(v.currentNodes) == "table" then for kk, vv in pairs(v.currentNodes) do if vv.currentNode == nil then CONFIG[k].currentNodes[kk] = nil end end else if v.currentNode == nil then if v.delete then v.delete() end CONFIG[k] = nil end end end end local function UrlEncode(szText) return szText:gsub("([^%w%-_%.%~])", function(c) return string.format("%%%02X", string.byte(c)) end) end local function UrlDecode(szText) return szText and szText:gsub("+", " "):gsub("%%(%x%x)", function(h) return string.char(tonumber(h, 16)) end) or nil end -- 取机场信息(剩余流量、到期时间) local subscribe_info = {} local function get_subscribe_info(cfgid, value) if type(cfgid) ~= "string" or cfgid == "" or type(value) ~= "string" then return end value = value:gsub("%s+", "") local expired_date = value:match("套餐到期:(.+)") local rem_traffic = value:match("剩余流量:(.+)") subscribe_info[cfgid] = subscribe_info[cfgid] or {expired_date = "", rem_traffic = ""} if expired_date then subscribe_info[cfgid]["expired_date"] = expired_date end if rem_traffic then subscribe_info[cfgid]["rem_traffic"] = rem_traffic end end -- 处理数据 local function processData(szType, content, add_mode, add_from) --log(content, add_mode, add_from) local result = { timeout = 60, add_mode = add_mode, --0为手动配置,1为导入,2为订阅 add_from = add_from } --ssr://base64(host:port:protocol:method:obfs:base64pass/?obfsparam=base64param&protoparam=base64param&remarks=base64remarks&group=base64group&udpport=0&uot=0) if szType == 'ssr' then if not has_ssr then log("跳过 SSR 节点,因未安装 SSR 核心程序 shadowsocksr-libev。") return nil end result.type = "SSR" local dat = split(content, "/%?") local hostInfo = split(dat[1], ':') if dat[1]:match('%[(.*)%]') then result.address = dat[1]:match('%[(.*)%]') else result.address = hostInfo[#hostInfo-5] end result.port = hostInfo[#hostInfo-4] result.protocol = hostInfo[#hostInfo-3] result.method = hostInfo[#hostInfo-2] result.obfs = hostInfo[#hostInfo-1] result.password = base64Decode(hostInfo[#hostInfo]) local params = {} for _, v in pairs(split(dat[2], '&')) do local t = split(v, '=') params[t[1]] = t[2] end result.obfs_param = base64Decode(params.obfsparam) result.protocol_param = base64Decode(params.protoparam) local group = base64Decode(params.group) if group then result.group = group end result.remarks = base64Decode(params.remarks) elseif szType == 'vmess' then local info = jsonParse(content) if vmess_type_default == "sing-box" and has_singbox then result.type = 'sing-box' elseif vmess_type_default == "xray" and has_xray then result.type = "Xray" else log("跳过 VMess 节点,因未适配到 VMess 核心程序,或未正确设置节点使用类型。") return nil end result.alter_id = info.aid result.address = info.add result.port = info.port result.protocol = 'vmess' result.uuid = info.id result.remarks = info.ps -- result.mux = 1 -- result.mux_concurrency = 8 if not info.net then info.net = "tcp" end info.net = string.lower(info.net) if result.type == "sing-box" and info.net == "raw" then info.net = "tcp" elseif result.type == "Xray" and info.net == "tcp" then info.net = "raw" end if info.net == 'h2' or info.net == 'http' then info.net = "http" result.transport = (result.type == "Xray") and "xhttp" or "http" else result.transport = info.net end if info.net == 'ws' then result.ws_host = info.host result.ws_path = info.path if result.type == "sing-box" and info.path then local ws_path_dat = split(info.path, "?") local ws_path = ws_path_dat[1] local ws_path_params = {} for _, v in pairs(split(ws_path_dat[2], '&')) do local t = split(v, '=') ws_path_params[t[1]] = t[2] end if ws_path_params.ed and tonumber(ws_path_params.ed) then result.ws_path = ws_path result.ws_enableEarlyData = "1" result.ws_maxEarlyData = tonumber(ws_path_params.ed) result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol" end end end if info.net == "http" then if result.type == "Xray" then result.xhttp_mode = "stream-one" result.xhttp_host = info.host result.xhttp_path = info.path else result.http_host = (info.host and info.host ~= "") and { info.host } or nil result.http_path = info.path end end if info.net == 'raw' or info.net == 'tcp' then if info.type and info.type ~= "http" then info.type = "none" end result.tcp_guise = info.type result.tcp_guise_http_host = (info.host and info.host ~= "") and { info.host } or nil result.tcp_guise_http_path = (info.path and info.path ~= "") and { info.path } or nil end if info.net == 'kcp' or info.net == 'mkcp' then info.net = "mkcp" result.mkcp_guise = info.type result.mkcp_mtu = 1350 result.mkcp_tti = 50 result.mkcp_uplinkCapacity = 5 result.mkcp_downlinkCapacity = 20 result.mkcp_readBufferSize = 2 result.mkcp_writeBufferSize = 2 result.mkcp_seed = info.seed end if info.net == 'quic' then result.quic_guise = info.type result.quic_key = info.key result.quic_security = info.securty end if info.net == 'grpc' then result.grpc_serviceName = info.path end if info.net == 'xhttp' then result.xhttp_host = info.host result.xhttp_path = info.path end if info.net == 'httpupgrade' then result.httpupgrade_host = info.host result.httpupgrade_path = info.path end if not info.security then result.security = "auto" end if info.tls == "tls" or info.tls == "1" then result.tls = "1" result.tls_serverName = (info.sni and info.sni ~= "") and info.sni or info.host info.allowinsecure = info.allowinsecure or info.insecure if info.allowinsecure and (info.allowinsecure == "1" or info.allowinsecure == "0") then result.tls_allowInsecure = info.allowinsecure else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end else result.tls = "0" end if result.type == "sing-box" and (result.transport == "mkcp" or result.transport == "xhttp") then log("跳过节点:" .. result.remarks ..",因Sing-Box不支持" .. szType .. "协议的" .. result.transport .. "传输方式,需更换Xray。") return nil end elseif szType == "ss" then if ss_type_default == "shadowsocks-libev" and has_ss then result.type = "SS" elseif ss_type_default == "shadowsocks-rust" and has_ss_rust then result.type = 'SS-Rust' elseif ss_type_default == "xray" and has_xray then result.type = 'Xray' result.protocol = 'shadowsocks' result.transport = 'raw' elseif ss_type_default == "sing-box" and has_singbox then result.type = 'sing-box' result.protocol = 'shadowsocks' else log("跳过 SS 节点,因未适配到 SS 核心程序,或未正确设置节点使用类型。") return nil end --SS-URI = "ss://" userinfo "@" hostname ":" port [ "/" ] [ "?" plugin ] [ "#" tag ] --userinfo = websafe-base64-encode-utf8(method ":" password) --ss://YWVzLTEyOC1nY206dGVzdA@192.168.100.1:8888#Example1 --ss://cmM0LW1kNTpwYXNzd2Q@192.168.100.1:8888/?plugin=obfs-local%3Bobfs%3Dhttp#Example2 --ss://2022-blake3-aes-256-gcm:YctPZ6U7xPPcU%2Bgp3u%2B0tx%2FtRizJN9K8y%2BuKlW2qjlI%3D@192.168.100.1:8888#Example3 --ss://2022-blake3-aes-256-gcm:YctPZ6U7xPPcU%2Bgp3u%2B0tx%2FtRizJN9K8y%2BuKlW2qjlI%3D@192.168.100.1:8888/?plugin=v2ray-plugin%3Bserver#Example3 --ss://Y2hhY2hhMjAtaWV0Zi1wb2x5MTMwNTp0ZXN0@xxxxxx.com:443?type=ws&path=%2Ftestpath&host=xxxxxx.com&security=tls&fp=&alpn=h3%2Ch2%2Chttp%2F1.1&sni=xxxxxx.com#test-1%40ss --ss://Y2hhY2hhMjAtaWV0Zi1wb2x5MTMwNTp4eHh4eHhAeHh4eC54eHh4eC5jb206NTYwMDE#Hong%20Kong-01 local idx_sp = content:find("#") or 0 local alias = "" if idx_sp > 0 then alias = content:sub(idx_sp + 1, -1) end result.remarks = UrlDecode(alias) local info = content:sub(1, idx_sp - 1):gsub("/%?", "?") local params = {} if info:find("%?") then local find_index = info:find("%?") local query = split(info, "%?") for _, v in pairs(split(query[2], '&')) do local t = split(v, '=') if #t >= 2 then params[t[1]] = UrlDecode(t[2]) end end if params.plugin then local plugin_info = params.plugin local idx_pn = plugin_info:find(";") if idx_pn then result.plugin = plugin_info:sub(1, idx_pn - 1) result.plugin_opts = plugin_info:sub(idx_pn + 1, #plugin_info) else result.plugin = plugin_info end end if result.plugin and result.plugin == "simple-obfs" then result.plugin = "obfs-local" end info = info:sub(1, find_index - 1) end local hostInfo = split(base64Decode(UrlDecode(info)), "@") if hostInfo and #hostInfo > 0 then local host_port = hostInfo[#hostInfo] -- [2001:4860:4860::8888]:443 -- 8.8.8.8:443 if host_port:find(":") then local sp = split(host_port, ":") result.port = sp[#sp] if api.is_ipv6addrport(host_port) then result.address = api.get_ipv6_only(host_port) else result.address = sp[1] end else result.address = host_port end local userinfo = nil if #hostInfo > 2 then userinfo = {} for i = 1, #hostInfo - 1 do tinsert(userinfo, hostInfo[i]) end userinfo = table.concat(userinfo, '@') else userinfo = base64Decode(hostInfo[1]) end local method, password if userinfo:find(":") then method = userinfo:sub(1, userinfo:find(":") - 1) password = userinfo:sub(userinfo:find(":") + 1, #userinfo) else password = hostInfo[1] --一些链接用明文uuid做密码 end -- 判断密码是否经过url编码 local function isURLEncodedPassword(pwd) if not pwd:find("%%[0-9A-Fa-f][0-9A-Fa-f]") then return false end local ok, decoded = pcall(UrlDecode, pwd) return ok and UrlEncode(decoded) == pwd end local decoded = UrlDecode(password) if isURLEncodedPassword(password) and decoded then password = decoded end local _method = (method or "none"):lower() method = (_method == "chacha20-poly1305" and "chacha20-ietf-poly1305") or (_method == "xchacha20-poly1305" and "xchacha20-ietf-poly1305") or _method result.method = method result.password = password if has_xray and (result.type ~= 'Xray' and result.type ~= 'sing-box' and params.type) then result.type = 'Xray' result.protocol = 'shadowsocks' elseif has_singbox and (result.type ~= 'Xray' and result.type ~= 'sing-box' and params.type) then result.type = 'sing-box' result.protocol = 'shadowsocks' end if result.plugin then if result.type == 'Xray' then -- obfs-local插件转换成xray支持的格式 if result.plugin ~= "obfs-local" then result.error_msg = "Xray不支持 " .. result.plugin .. " 插件." else local obfs = result.plugin_opts:match("obfs=([^;]+)") or "" local obfs_host = result.plugin_opts:match("obfs%-host=([^;]+)") or "" if obfs == "" or obfs_host == "" then result.error_msg = "SS " .. result.plugin .. " 插件选项不完整." end if obfs == "http" then result.transport = "raw" result.tcp_guise = "http" result.tcp_guise_http_host = (obfs_host and obfs_host ~= "") and { obfs_host } or nil result.tcp_guise_http_path = { "/" } elseif obfs == "tls" then result.tls = "1" result.tls_serverName = obfs_host result.tls_allowInsecure = "1" end result.plugin = nil result.plugin_opts = nil end else result.plugin_enabled = "1" end end if result.type == "SS" then local aead2022_methods = { "2022-blake3-aes-128-gcm", "2022-blake3-aes-256-gcm", "2022-blake3-chacha20-poly1305" } local aead2022 = false for k, v in ipairs(aead2022_methods) do if method:lower() == v:lower() then aead2022 = true end end if aead2022 then -- shadowsocks-libev 不支持2022加密 result.error_msg = "shadowsocks-libev 不支持2022加密." end end if params.type then params.type = string.lower(params.type) if result.type == "sing-box" and params.type == "raw" then params.type = "tcp" elseif result.type == "Xray" and params.type == "tcp" then params.type = "raw" end if params.type == "h2" or params.type == "http" then params.type = "http" result.transport = (result.type == "Xray") and "xhttp" or "http" else result.transport = params.type end if result.type ~= "SS-Rust" and result.type ~= "SS" then if params.type == 'ws' then result.ws_host = params.host result.ws_path = params.path if result.type == "sing-box" and params.path then local ws_path_dat = split(params.path, "%?") local ws_path = ws_path_dat[1] local ws_path_params = {} for _, v in pairs(split(ws_path_dat[2], '&')) do local t = split(v, '=') ws_path_params[t[1]] = t[2] end if ws_path_params.ed and tonumber(ws_path_params.ed) then result.ws_path = ws_path result.ws_enableEarlyData = "1" result.ws_maxEarlyData = tonumber(ws_path_params.ed) result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol" end end end if params.type == "http" then if result.type == "sing-box" then result.transport = "http" result.http_host = (params.host and params.host ~= "") and { params.host } or nil result.http_path = params.path elseif result.type == "Xray" then result.transport = "xhttp" result.xhttp_mode = "stream-one" result.xhttp_host = params.host result.xhttp_path = params.path end end if params.type == 'raw' or params.type == 'tcp' then result.tcp_guise = params.headerType or "none" result.tcp_guise_http_host = (params.host and params.host ~= "") and { params.host } or nil result.tcp_guise_http_path = (params.path and params.path ~= "") and { params.path } or nil end if params.type == 'kcp' or params.type == 'mkcp' then result.transport = "mkcp" result.mkcp_guise = params.headerType or "none" result.mkcp_mtu = 1350 result.mkcp_tti = 50 result.mkcp_uplinkCapacity = 5 result.mkcp_downlinkCapacity = 20 result.mkcp_readBufferSize = 2 result.mkcp_writeBufferSize = 2 result.mkcp_seed = params.seed end if params.type == 'quic' then result.quic_guise = params.headerType or "none" result.quic_key = params.key result.quic_security = params.quicSecurity or "none" end if params.type == 'grpc' then if params.path then result.grpc_serviceName = params.path end if params.serviceName then result.grpc_serviceName = params.serviceName end result.grpc_mode = params.mode or "gun" end result.tls = "0" if params.security == "tls" or params.security == "reality" then result.tls = "1" result.tls_serverName = (params.sni and params.sni ~= "") and params.sni or params.host result.alpn = params.alpn if params.fp and params.fp ~= "" then result.utls = "1" result.fingerprint = params.fp end if params.ech and params.ech ~= "" then result.ech = "1" result.ech_config = params.ech end if params.security == "reality" then result.reality = "1" result.reality_publicKey = params.pbk or nil result.reality_shortId = params.sid or nil result.reality_spiderX = params.spx or nil result.use_mldsa65Verify = (params.pqv and params.pqv ~= "") and "1" or nil result.reality_mldsa65Verify = params.pqv or nil end end params.allowinsecure = params.allowinsecure or params.insecure if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then result.tls_allowInsecure = params.allowinsecure else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end else result.error_msg = "请更换Xray或Sing-Box来支持SS更多的传输方式." end end if params["shadow-tls"] then if result.type ~= "sing-box" and result.type ~= "SS-Rust" then result.error_msg = ss_type_default .. " 不支持 shadow-tls 插件." else -- 解析SS Shadow-TLS 插件参数 local function parseShadowTLSParams(b64str, out) local ok, data = pcall(jsonParse, base64Decode(b64str)) if not ok or type(data) ~= "table" then return "" end if type(out) == "table" then for k, v in pairs(data) do out[k] = v end end local t = {} if data.version then t[#t+1] = "v" .. data.version .. "=1" end if data.password then t[#t+1] = "passwd=" .. data.password end for k, v in pairs(data) do if k ~= "version" and k ~= "password" then t[#t+1] = k .. "=" .. tostring(v) end end return table.concat(t, ";") end if result.type == "SS-Rust" then result.plugin_enabled = "1" result.plugin = "shadow-tls" result.plugin_opts = parseShadowTLSParams(params["shadow-tls"]) elseif result.type == "sing-box" then local shadowtlsOpt = {} parseShadowTLSParams(params["shadow-tls"], shadowtlsOpt) if next(shadowtlsOpt) then result.shadowtls = "1" result.shadowtls_version = shadowtlsOpt.version or "1" result.shadowtls_password = shadowtlsOpt.password result.shadowtls_serverName = shadowtlsOpt.host if shadowtlsOpt.fingerprint then result.shadowtls_utls = "1" result.shadowtls_fingerprint = shadowtlsOpt.fingerprint or "chrome" end end end end end end elseif szType == "trojan" then if trojan_type_default == "trojan-plus" and has_trojan_plus then result.type = "Trojan-Plus" elseif trojan_type_default == "sing-box" and has_singbox then result.type = 'sing-box' result.protocol = 'trojan' elseif trojan_type_default == "xray" and has_xray then result.type = 'Xray' result.protocol = 'trojan' else log("跳过 Trojan 节点,因未适配到 Trojan 核心程序,或未正确设置节点使用类型。") return nil end local alias = "" if content:find("#") then local idx_sp = content:find("#") alias = content:sub(idx_sp + 1, -1) content = content:sub(0, idx_sp - 1) end result.remarks = UrlDecode(alias) if content:find("@") then local Info = split(content, "@") result.password = UrlDecode(Info[1]) local port = "443" Info[2] = (Info[2] or ""):gsub("/%?", "?") local query = split(Info[2], "%?") local host_port = query[1] local params = {} for _, v in pairs(split(query[2], '&')) do local t = split(v, '=') if #t > 1 then params[string.lower(t[1])] = UrlDecode(t[2]) end end -- [2001:4860:4860::8888]:443 -- 8.8.8.8:443 if host_port:find(":") then local sp = split(host_port, ":") port = sp[#sp] if api.is_ipv6addrport(host_port) then result.address = api.get_ipv6_only(host_port) else result.address = sp[1] end else result.address = host_port end local peer, sni = nil, "" if params.peer then peer = params.peer end sni = params.sni and params.sni or "" result.port = port result.tls = '1' result.tls_serverName = peer and peer or sni params.allowinsecure = params.allowinsecure or params.insecure if params.allowinsecure then if params.allowinsecure == "1" or params.allowinsecure == "0" then result.tls_allowInsecure = params.allowinsecure else result.tls_allowInsecure = string.lower(params.allowinsecure) == "true" and "1" or "0" end --log(result.remarks .. ' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure) else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end if not params.type then params.type = "tcp" end params.type = string.lower(params.type) if result.type == "sing-box" and params.type == "raw" then params.type = "tcp" elseif result.type == "Xray" and params.type == "tcp" then params.type = "raw" end if params.type == "h2" or params.type == "http" then params.type = "http" result.transport = (result.type == "Xray") and "xhttp" or "http" else result.transport = params.type end if params.type == 'ws' then result.ws_host = params.host result.ws_path = params.path if result.type == "sing-box" and params.path then local ws_path_dat = split(params.path, "%?") local ws_path = ws_path_dat[1] local ws_path_params = {} for _, v in pairs(split(ws_path_dat[2], '&')) do local t = split(v, '=') ws_path_params[t[1]] = t[2] end if ws_path_params.ed and tonumber(ws_path_params.ed) then result.ws_path = ws_path result.ws_enableEarlyData = "1" result.ws_maxEarlyData = tonumber(ws_path_params.ed) result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol" end end end if params.type == "http" then if result.type == "sing-box" then result.transport = "http" result.http_host = (params.host and params.host ~= "") and { params.host } or nil result.http_path = params.path elseif result.type == "Xray" then result.transport = "xhttp" result.xhttp_mode = "stream-one" result.xhttp_host = params.host result.xhttp_path = params.path end end if params.type == 'raw' or params.type == 'tcp' then result.tcp_guise = params.headerType or "none" result.tcp_guise_http_host = (params.host and params.host ~= "") and { params.host } or nil result.tcp_guise_http_path = (params.path and params.path ~= "") and { params.path } or nil end if params.type == 'kcp' or params.type == 'mkcp' then result.transport = "mkcp" result.mkcp_guise = params.headerType or "none" result.mkcp_mtu = 1350 result.mkcp_tti = 50 result.mkcp_uplinkCapacity = 5 result.mkcp_downlinkCapacity = 20 result.mkcp_readBufferSize = 2 result.mkcp_writeBufferSize = 2 result.mkcp_seed = params.seed end if params.type == 'quic' then result.quic_guise = params.headerType or "none" result.quic_key = params.key result.quic_security = params.quicSecurity or "none" end if params.type == 'grpc' then if params.path then result.grpc_serviceName = params.path end if params.serviceName then result.grpc_serviceName = params.serviceName end result.grpc_mode = params.mode or "gun" end if params.type == 'xhttp' then result.xhttp_host = params.host result.xhttp_path = params.path end if params.type == 'httpupgrade' then result.httpupgrade_host = params.host result.httpupgrade_path = params.path end result.alpn = params.alpn if result.type == "sing-box" and (result.transport == "mkcp" or result.transport == "xhttp") then log("跳过节点:" .. result.remarks ..",因Sing-Box不支持" .. szType .. "协议的" .. result.transport .. "传输方式,需更换Xray。") return nil end end elseif szType == "ssd" then result.type = "SS" result.address = content.server result.port = content.port result.password = content.password result.method = content.encryption result.plugin = content.plugin result.plugin_opts = content.plugin_options result.group = content.airport result.remarks = content.remarks elseif szType == "vless" then if vless_type_default == "sing-box" and has_singbox then result.type = 'sing-box' elseif vless_type_default == "xray" and has_xray then result.type = "Xray" else log("跳过 VLESS 节点,因未适配到 VLESS 核心程序,或未正确设置节点使用类型。") return nil end result.protocol = "vless" local alias = "" if content:find("#") then local idx_sp = content:find("#") alias = content:sub(idx_sp + 1, -1) content = content:sub(0, idx_sp - 1) end result.remarks = UrlDecode(alias) if content:find("@") then local Info = split(content, "@") result.uuid = UrlDecode(Info[1]) local port = "443" Info[2] = (Info[2] or ""):gsub("/%?", "?") local query = split(Info[2], "%?") local host_port = query[1] local params = {} for _, v in pairs(split(query[2], '&')) do local t = split(v, '=') params[t[1]] = UrlDecode(t[2]) end -- [2001:4860:4860::8888]:443 -- 8.8.8.8:443 if host_port:find(":") then local sp = split(host_port, ":") port = sp[#sp] if api.is_ipv6addrport(host_port) then result.address = api.get_ipv6_only(host_port) else result.address = sp[1] end else result.address = host_port end if not params.type then params.type = "tcp" end params.type = string.lower(params.type) if ({ xhttp=true, kcp=true, mkcp=true })[params.type] and result.type ~= "Xray" and has_xray then result.type = "Xray" end if result.type == "sing-box" and params.type == "raw" then params.type = "tcp" elseif result.type == "Xray" and params.type == "tcp" then params.type = "raw" end if params.type == "h2" or params.type == "http" then params.type = "http" result.transport = (result.type == "Xray") and "xhttp" or "http" else result.transport = params.type end if params.type == 'ws' then result.ws_host = params.host result.ws_path = params.path if result.type == "sing-box" and params.path then local ws_path_dat = split(params.path, "%?") local ws_path = ws_path_dat[1] local ws_path_params = {} for _, v in pairs(split(ws_path_dat[2], '&')) do local t = split(v, '=') ws_path_params[t[1]] = t[2] end if ws_path_params.ed and tonumber(ws_path_params.ed) then result.ws_path = ws_path result.ws_enableEarlyData = "1" result.ws_maxEarlyData = tonumber(ws_path_params.ed) result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol" end end end if params.type == "http" then if result.type == "sing-box" then result.transport = "http" result.http_host = (params.host and params.host ~= "") and { params.host } or nil result.http_path = params.path elseif result.type == "Xray" then result.transport = "xhttp" result.xhttp_mode = "stream-one" result.xhttp_host = params.host result.xhttp_path = params.path end end if params.type == 'raw' or params.type == 'tcp' then result.tcp_guise = params.headerType or "none" result.tcp_guise_http_host = (params.host and params.host ~= "") and { params.host } or nil result.tcp_guise_http_path = (params.path and params.path ~= "") and { params.path } or nil end if params.type == 'kcp' or params.type == 'mkcp' then result.transport = "mkcp" result.mkcp_guise = params.headerType or "none" result.mkcp_mtu = 1350 result.mkcp_tti = 50 result.mkcp_uplinkCapacity = 5 result.mkcp_downlinkCapacity = 20 result.mkcp_readBufferSize = 2 result.mkcp_writeBufferSize = 2 result.mkcp_seed = params.seed end if params.type == 'quic' then result.quic_guise = params.headerType or "none" result.quic_key = params.key result.quic_security = params.quicSecurity or "none" end if params.type == 'grpc' then if params.path then result.grpc_serviceName = params.path end if params.serviceName then result.grpc_serviceName = params.serviceName end result.grpc_mode = params.mode or "gun" end if params.type == 'xhttp' then result.xhttp_host = params.host result.xhttp_path = params.path result.xhttp_mode = params.mode or "auto" result.use_xhttp_extra = (params.extra and params.extra ~= "") and "1" or nil result.xhttp_extra = (params.extra and params.extra ~= "") and params.extra or nil local success, Data = pcall(jsonParse, params.extra) if success and Data then local address = (Data.extra and Data.extra.downloadSettings and Data.extra.downloadSettings.address) or (Data.downloadSettings and Data.downloadSettings.address) result.download_address = (address and address ~= "") and address:gsub("^%[", ""):gsub("%]$", "") or nil else result.download_address = nil end end if params.type == 'httpupgrade' then result.httpupgrade_host = params.host result.httpupgrade_path = params.path end result.encryption = params.encryption or "none" result.flow = params.flow and params.flow:gsub("-udp443", "") or nil result.tls = "0" if params.security == "tls" or params.security == "reality" then result.tls = "1" result.tls_serverName = (params.sni and params.sni ~= "") and params.sni or params.host result.alpn = params.alpn if params.fp and params.fp ~= "" then result.utls = "1" result.fingerprint = params.fp end if params.ech and params.ech ~= "" then result.ech = "1" result.ech_config = params.ech end if params.security == "reality" then result.reality = "1" result.reality_publicKey = params.pbk or nil result.reality_shortId = params.sid or nil result.reality_spiderX = params.spx or nil result.use_mldsa65Verify = (params.pqv and params.pqv ~= "") and "1" or nil result.reality_mldsa65Verify = params.pqv or nil end end result.port = port params.allowinsecure = params.allowinsecure or params.insecure if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then result.tls_allowInsecure = params.allowinsecure else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end if result.type == "sing-box" and (result.transport == "mkcp" or result.transport == "xhttp") then log("跳过节点:" .. result.remarks ..",因Sing-Box不支持" .. szType .. "协议的" .. result.transport .. "传输方式,需更换Xray。") return nil end end elseif szType == 'hysteria' then if has_singbox then result.type = 'sing-box' result.protocol = "hysteria" else log("跳过 Hysteria 节点,因未安装 Hysteria 核心程序 Sing-box。") return nil end local alias = "" if content:find("#") then local idx_sp = content:find("#") alias = content:sub(idx_sp + 1, -1) content = content:sub(0, idx_sp - 1) end result.remarks = UrlDecode(alias) local dat = split(content:gsub("/%?", "?"), '%?') local host_port = dat[1] local params = {} for _, v in pairs(split(dat[2], '&')) do local t = split(v, '=') if #t > 0 then params[t[1]] = t[2] end end -- [2001:4860:4860::8888]:443 -- 8.8.8.8:443 if host_port:find(":") then local sp = split(host_port, ":") result.port = sp[#sp] if api.is_ipv6addrport(host_port) then result.address = api.get_ipv6_only(host_port) else result.address = sp[1] end else result.address = host_port end result.hysteria_obfs = params.obfsParam result.hysteria_auth_type = "string" result.hysteria_auth_password = params.auth result.tls_serverName = params.peer params.allowinsecure = params.allowinsecure or params.insecure if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then result.tls_allowInsecure = params.allowinsecure --log(result.remarks ..' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure) else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end result.hysteria_alpn = params.alpn result.hysteria_up_mbps = params.upmbps result.hysteria_down_mbps = params.downmbps result.hysteria_hop = params.mport elseif szType == 'hysteria2' or szType == 'hy2' then local alias = "" if content:find("#") then local idx_sp = content:find("#") alias = content:sub(idx_sp + 1, -1) content = content:sub(0, idx_sp - 1) end result.remarks = UrlDecode(alias) local Info = content if content:find("@") then local contents = split(content, "@") result.hysteria2_auth_password = UrlDecode(contents[1]) Info = (contents[2] or ""):gsub("/%?", "?") end local query = split(Info, "%?") local host_port = query[1] local params = {} for _, v in pairs(split(query[2], '&')) do local t = split(v, '=') if #t > 1 then params[string.lower(t[1])] = UrlDecode(t[2]) end end -- [2001:4860:4860::8888]:443 -- 8.8.8.8:443 if host_port:find(":") then local sp = split(host_port, ":") result.port = sp[#sp] if api.is_ipv6addrport(host_port) then result.address = api.get_ipv6_only(host_port) else result.address = sp[1] end else result.address = host_port end result.tls_serverName = params.sni params.allowinsecure = params.allowinsecure or params.insecure if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then result.tls_allowInsecure = params.allowinsecure --log(result.remarks ..' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure) else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end result.hysteria2_tls_pinSHA256 = params.pinSHA256 result.hysteria2_hop = params.mport if hysteria2_type_default == "sing-box" and has_singbox then result.type = 'sing-box' result.protocol = "hysteria2" if params["obfs-password"] or params["obfs_password"] then result.hysteria2_obfs_type = "salamander" result.hysteria2_obfs_password = params["obfs-password"] or params["obfs_password"] end elseif has_hysteria2 then result.type = "Hysteria2" if params["obfs-password"] or params["obfs_password"] then result.hysteria2_obfs = params["obfs-password"] or params["obfs_password"] end else log("跳过 Hysteria2 节点,因未适配到 Hysteria2 核心程序,或未正确设置节点使用类型。") return nil end elseif szType == 'tuic' then if has_singbox then result.type = 'sing-box' result.protocol = "tuic" else log("跳过 Tuic 节点,因未安装 Tuic 核心程序 Sing-box。") return nil end local alias = "" if content:find("#") then local idx_sp = content:find("#") alias = content:sub(idx_sp + 1, -1) content = content:sub(0, idx_sp - 1) end result.remarks = UrlDecode(alias) local Info = content if content:find("@") then local contents = split(content, "@") if contents[1]:find(":") then local userinfo = split(contents[1], ":") result.uuid = UrlDecode(userinfo[1]) result.password = UrlDecode(userinfo[2]) end Info = (contents[2] or ""):gsub("/%?", "?") end local query = split(Info, "%?") local host_port = query[1] local params = {} for _, v in pairs(split(query[2], '&')) do local t = split(v, '=') if #t > 1 then params[string.lower(t[1])] = UrlDecode(t[2]) end end if host_port:find(":") then local sp = split(host_port, ":") result.port = sp[#sp] if api.is_ipv6addrport(host_port) then result.address = api.get_ipv6_only(host_port) else result.address = sp[1] end else result.address = host_port end result.tls_serverName = params.sni result.tuic_alpn = params.alpn or "default" result.tuic_congestion_control = params.congestion_control or "cubic" result.tuic_udp_relay_mode = params.udp_relay_mode or "native" params.allowinsecure = params.allowinsecure or params.insecure if params.allowinsecure then if params.allowinsecure == "1" or params.allowinsecure == "0" then result.tls_allowInsecure = params.allowinsecure else result.tls_allowInsecure = string.lower(params.allowinsecure) == "true" and "1" or "0" end --log(result.remarks .. ' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure) else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end elseif szType == "anytls" then if has_singbox then result.type = 'sing-box' result.protocol = "anytls" else log("跳过 AnyTLS 节点,因未安装 AnyTLS 核心程序 Sing-box 1.12。") return nil end local alias = "" if content:find("#") then local idx_sp = content:find("#") alias = content:sub(idx_sp + 1, -1) content = content:sub(0, idx_sp - 1) end result.remarks = UrlDecode(alias) if content:find("@") then local Info = split(content, "@") result.password = UrlDecode(Info[1]) local port = "443" Info[2] = (Info[2] or ""):gsub("/%?", "?") local query = split(Info[2], "%?") local host_port = query[1] local params = {} for _, v in pairs(split(query[2], '&')) do local t = split(v, '=') params[t[1]] = UrlDecode(t[2]) end -- [2001:4860:4860::8888]:443 -- 8.8.8.8:443 if host_port:find(":") then local sp = split(host_port, ":") port = sp[#sp] if api.is_ipv6addrport(host_port) then result.address = api.get_ipv6_only(host_port) else result.address = sp[1] end else result.address = host_port end result.tls = "0" if (not params.security or params.security == "") and params.sni and params.sni ~= "" then params.security = "tls" end if params.security == "tls" or params.security == "reality" then result.tls = "1" result.tls_serverName = params.sni result.alpn = params.alpn if params.fp and params.fp ~= "" then result.utls = "1" result.fingerprint = params.fp end if params.security == "reality" then result.reality = "1" result.reality_publicKey = params.pbk or nil result.reality_shortId = params.sid or nil end end result.port = port params.allowinsecure = params.allowinsecure or params.insecure if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then result.tls_allowInsecure = params.allowinsecure else result.tls_allowInsecure = allowInsecure_default and "1" or "0" end local singbox_version = api.get_app_version("sing-box") local version_ge_1_12 = api.compare_versions(singbox_version:match("[^v]+"), ">=", "1.12.0") if not has_singbox or not version_ge_1_12 then log("跳过节点:" .. result.remarks ..",因" .. szType .. "类型的节点需要 Sing-Box 1.12 以上版本支持。") return nil end end else log('暂时不支持' .. szType .. "类型的节点订阅,跳过此节点。") return nil end if not result.remarks or result.remarks == "" then if result.address and result.port then result.remarks = result.address .. ':' .. result.port else result.remarks = "NULL" end end return result end local function curl(url, file, ua, mode) local curl_args = { "-skL", "-w %{http_code}", "--retry 3", "--connect-timeout 3" } if ua and ua ~= "" and ua ~= "curl" then curl_args[#curl_args + 1] = '--user-agent "' .. ua .. '"' end local return_code, result if mode == "direct" then return_code, result = api.curl_direct(url, file, curl_args) elseif mode == "proxy" then return_code, result = api.curl_proxy(url, file, curl_args) else return_code, result = api.curl_auto(url, file, curl_args) end return tonumber(result) end local function truncate_nodes(add_from) for _, config in pairs(CONFIG) do if config.currentNodes and #config.currentNodes > 0 then local newNodes = {} local removeNodesSet = {} for k, v in pairs(config.currentNodes) do if v.currentNode and v.currentNode.add_mode == "2" then if (not add_from) or (add_from and add_from == v.currentNode.add_from) then removeNodesSet[v.currentNode[".name"]] = true end end end for _, value in ipairs(config.currentNodes) do if not removeNodesSet[value.currentNode[".name"]] then newNodes[#newNodes + 1] = value.currentNode[".name"] end end if config.set then config.set(config, newNodes) end else if config.currentNode and config.currentNode.add_mode == "2" then if (not add_from) or (add_from and add_from == config.currentNode.add_from) then if config.delete then config.delete(config) elseif config.set then config.set(config, "") end end end end end uci:foreach(appname, "nodes", function(node) if node.add_mode == "2" then if (not add_from) or (add_from and add_from == node.add_from) then uci:delete(appname, node['.name']) end end end) uci:foreach(appname, "subscribe_list", function(o) if (not add_from) or add_from == o.remark then uci:delete(appname, o['.name'], "md5") end end) api.uci_save(uci, appname, true) end local function select_node(nodes, config, parentConfig) if config.currentNode then local server -- 特别优先级 cfgid if config.currentNode[".name"] then for index, node in pairs(nodes) do if node[".name"] == config.currentNode[".name"] then if config.log == nil or config.log == true then log('更新【' .. config.remarks .. '】匹配节点:' .. node.remarks) end server = node[".name"] break end end end -- 第一优先级 类型 + 备注 + IP + 端口 if not server then for index, node in pairs(nodes) do if config.currentNode.type and config.currentNode.remarks and config.currentNode.address and config.currentNode.port then if node.type and node.remarks and node.address and node.port then if node.type == config.currentNode.type and node.remarks == config.currentNode.remarks and (node.address .. ':' .. node.port == config.currentNode.address .. ':' .. config.currentNode.port) then if config.log == nil or config.log == true then log('更新【' .. config.remarks .. '】第一匹配节点:' .. node.remarks) end server = node[".name"] break end end end end end -- 第二优先级 类型 + IP + 端口 if not server then for index, node in pairs(nodes) do if config.currentNode.type and config.currentNode.address and config.currentNode.port then if node.type and node.address and node.port then if node.type == config.currentNode.type and (node.address .. ':' .. node.port == config.currentNode.address .. ':' .. config.currentNode.port) then if config.log == nil or config.log == true then log('更新【' .. config.remarks .. '】第二匹配节点:' .. node.remarks) end server = node[".name"] break end end end end end -- 第三优先级 IP + 端口 if not server then for index, node in pairs(nodes) do if config.currentNode.address and config.currentNode.port then if node.address and node.port then if node.address .. ':' .. node.port == config.currentNode.address .. ':' .. config.currentNode.port then if config.log == nil or config.log == true then log('更新【' .. config.remarks .. '】第三匹配节点:' .. node.remarks) end server = node[".name"] break end end end end end -- 第四优先级 IP if not server then for index, node in pairs(nodes) do if config.currentNode.address then if node.address then if node.address == config.currentNode.address then if config.log == nil or config.log == true then log('更新【' .. config.remarks .. '】第四匹配节点:' .. node.remarks) end server = node[".name"] break end end end end end -- 第五优先级备注 if not server then for index, node in pairs(nodes) do if config.currentNode.remarks then if node.remarks then if node.remarks == config.currentNode.remarks then if config.log == nil or config.log == true then log('更新【' .. config.remarks .. '】第五匹配节点:' .. node.remarks) end server = node[".name"] break end end end end end if not parentConfig then -- 还不行 随便找一个 if not server then if #nodes_table > 0 then if config.log == nil or config.log == true then log('【' .. config.remarks .. '】' .. '无法找到最匹配的节点,当前已更换为:' .. nodes_table[1].remarks) end server = nodes_table[1][".name"] end end end if server then if parentConfig then config.set(parentConfig, server) else config.set(config, server) end end else if not parentConfig then config.set(config, "") end end end local function update_node(manual) if next(nodeResult) == nil then log("没有可用的节点信息更新。") return end local group = {} for _, v in ipairs(nodeResult) do group[v["remark"]] = true end if manual == 0 and next(group) then uci:foreach(appname, "nodes", function(node) -- 如果未发现新节点或手动导入的节点就不要删除了... if node.add_mode == "2" and (node.add_from and group[node.add_from] == true) then uci:delete(appname, node['.name']) end end) end for _, v in ipairs(nodeResult) do local remark = v["remark"] local list = v["list"] for _, vv in ipairs(list) do local cfgid = uci:section(appname, "nodes", api.gen_short_uuid()) for kkk, vvv in pairs(vv) do if type(vvv) == "table" and next(vvv) ~= nil then uci:set_list(appname, cfgid, kkk, vvv) else uci:set(appname, cfgid, kkk, vvv) -- sing-box 域名解析策略 if kkk == "type" and vvv == "sing-box" then uci:set(appname, cfgid, "domain_strategy", domain_strategy_node) end -- 订阅组链式代理 if chain_node_type ~= "" and kkk == "type" and vvv == chain_node_type then if preproxy_node_group ~="" then uci:set(appname, cfgid, "chain_proxy", "1") uci:set(appname, cfgid, "preproxy_node", preproxy_node_group) elseif to_node_group ~= "" then uci:set(appname, cfgid, "chain_proxy", "2") uci:set(appname, cfgid, "to_node", to_node_group) end end end end end end -- 更新机场信息 for cfgid, info in pairs(subscribe_info) do for key, value in pairs(info) do if value ~= "" then uci:set(appname, cfgid, key, value) else uci:delete(appname, cfgid, key) end end end api.uci_save(uci, appname, true) if next(CONFIG) then local nodes = {} uci:foreach(appname, "nodes", function(node) nodes[#nodes + 1] = node end) for _, config in pairs(CONFIG) do if config.currentNodes and #config.currentNodes > 0 then if config.remarks and config.currentNodes[1].log ~= false then log('----【' .. config.remarks .. '】----') end for kk, vv in pairs(config.currentNodes) do select_node(nodes, vv, config) end config.set(config) else select_node(nodes, config) end end api.uci_save(uci, appname, true) end if arg[3] == "cron" then if not fs.access("/var/lock/" .. appname .. ".lock") then luci.sys.call("touch /tmp/lock/" .. appname .. "_cron.lock") end end luci.sys.call("/etc/init.d/" .. appname .. " restart > /dev/null 2>&1 &") end local function parse_link(raw, add_mode, add_from, cfgid) if raw and #raw > 0 then local nodes, szType local node_list = {} -- SSD 似乎是这种格式 ssd:// 开头的 if raw:find('ssd://') then szType = 'ssd' local nEnd = select(2, raw:find('ssd://')) nodes = base64Decode(raw:sub(nEnd + 1, #raw)) nodes = jsonParse(nodes) local extra = { airport = nodes.airport, port = nodes.port, encryption = nodes.encryption, password = nodes.password } local servers = {} -- SS里面包着 干脆直接这样 for _, server in ipairs(nodes.servers) do tinsert(servers, setmetatable(server, { __index = extra })) end nodes = servers else -- ssd 外的格式 if add_mode == "1" then nodes = split(raw, "\n") else nodes = split(base64Decode(raw):gsub("\r\n", "\n"), "\n") end end for _, v in ipairs(nodes) do if v and not string.match(v, "^%s*$") then xpcall(function () local result if szType == 'ssd' then result = processData(szType, v, add_mode, add_from) elseif not szType then local node = api.trim(v) local dat = split(node, "://") if dat and dat[1] and dat[2] then if dat[1] == 'vmess' or dat[1] == 'ssr' then local link = api.trim(dat[2]:gsub("#.*$", "")) result = processData(dat[1], base64Decode(link), add_mode, add_from) else local link = dat[2]:gsub("&amp;", "&"):gsub("%s*#%s*", "#") -- 一些奇葩的链接用"&amp;"当做"&","#"前后带空格 result = processData(dat[1], link, add_mode, add_from) end end else log('跳过未知类型: ' .. szType) end -- log(result) if result then if result.error_msg then log('丢弃节点: ' .. result.remarks .. ", 原因:" .. result.error_msg) elseif not result.type then log('丢弃节点: ' .. result.remarks .. ", 找不到可使用二进制.") elseif (add_mode == "2" and is_filter_keyword(result.remarks)) or not result.address or result.remarks == "NULL" or result.address == "127.0.0.1" or (not datatypes.hostname(result.address) and not (api.is_ip(result.address))) then log('丢弃过滤节点: ' .. result.type .. ' 节点, ' .. result.remarks) else tinsert(node_list, result) end if add_mode == "2" then get_subscribe_info(cfgid, result.remarks) end end end, function (err) --log(err) log(v, "解析错误,跳过此节点。") end ) end end if #node_list > 0 then nodeResult[#nodeResult + 1] = { remark = add_from, list = node_list } end log('成功解析【' .. add_from .. '】节点数量: ' .. #node_list) else if add_mode == "2" then log('获取到的【' .. add_from .. '】订阅内容为空,可能是订阅地址无效,或是网络问题,请诊断!') end end end local execute = function() do local subscribe_list = {} local fail_list = {} if arg[2] ~= "all" then string.gsub(arg[2], '[^' .. "," .. ']+', function(w) subscribe_list[#subscribe_list + 1] = uci:get_all(appname, w) or {} end) else uci:foreach(appname, "subscribe_list", function(o) subscribe_list[#subscribe_list + 1] = o end) end local manual_sub = arg[3] == "manual" for index, value in ipairs(subscribe_list) do local cfgid = value[".name"] local remark = value.remark local url = value.url if value.allowInsecure and value.allowInsecure == "1" then allowInsecure_default = true end local filter_keyword_mode = value.filter_keyword_mode or "5" if filter_keyword_mode == "0" then filter_keyword_mode_default = "0" elseif filter_keyword_mode == "1" then filter_keyword_mode_default = "1" filter_keyword_discard_list_default = value.filter_discard_list or {} elseif filter_keyword_mode == "2" then filter_keyword_mode_default = "2" filter_keyword_keep_list_default = value.filter_keep_list or {} elseif filter_keyword_mode == "3" then filter_keyword_mode_default = "3" filter_keyword_keep_list_default = value.filter_keep_list or {} filter_keyword_discard_list_default = value.filter_discard_list or {} elseif filter_keyword_mode == "4" then filter_keyword_mode_default = "4" filter_keyword_keep_list_default = value.filter_keep_list or {} filter_keyword_discard_list_default = value.filter_discard_list or {} end local ss_type = value.ss_type or "global" if ss_type ~= "global" then ss_type_default = ss_type end local trojan_type = value.trojan_type or "global" if trojan_type ~= "global" then trojan_type_default = trojan_type end local vmess_type = value.vmess_type or "global" if vmess_type ~= "global" then vmess_type_default = vmess_type end local vless_type = value.vless_type or "global" if vless_type ~= "global" then vless_type_default = vless_type end local hysteria2_type = value.hysteria2_type or "global" if hysteria2_type ~= "global" then hysteria2_type_default = hysteria2_type end local domain_strategy = value.domain_strategy or "global" if domain_strategy ~= "global" then domain_strategy_node = domain_strategy else domain_strategy_node = domain_strategy_default end -- 订阅组链式代理 local function valid_chain_node(node) if not node then return "" end local cp = uci:get(appname, node, "chain_proxy") or "" local am = uci:get(appname, node, "add_mode") or "0" chain_node_type = (cp == "" and am ~= "2") and (uci:get(appname, node, "type") or "") or "" if chain_node_type ~= "Xray" and chain_node_type ~= "sing-box" then chain_node_type = "" return "" end return node end preproxy_node_group = (value.chain_proxy == "1") and valid_chain_node(value.preproxy_node) or "" to_node_group = (value.chain_proxy == "2") and valid_chain_node(value.to_node) or "" local ua = value.user_agent local access_mode = value.access_mode local result = (not access_mode) and "自动" or (access_mode == "direct" and "直连访问" or (access_mode == "proxy" and "通过代理" or "自动")) log('正在订阅:【' .. remark .. '】' .. url .. ' [' .. result .. ']') local tmp_file = "/tmp/" .. cfgid value.http_code = curl(url, tmp_file, ua, access_mode) if value.http_code ~= 200 then fail_list[#fail_list + 1] = value else if luci.sys.call("[ -f " .. tmp_file .. " ] && sed -i -e '/^[ \t]*$/d' -e '/^[ \t]*\r$/d' " .. tmp_file) == 0 then local f = io.open(tmp_file, "r") local stdout = f:read("*all") f:close() local raw_data = api.trim(stdout) local old_md5 = value.md5 or "" local new_md5 = luci.sys.exec("md5sum " .. tmp_file .. " 2>/dev/null | awk '{print $1}'"):gsub("\n", "") if not manual_sub and old_md5 == new_md5 then log('订阅:【' .. remark .. '】没有变化,无需更新。') else parse_link(raw_data, "2", remark, cfgid) uci:set(appname, cfgid, "md5", new_md5) end else fail_list[#fail_list + 1] = value end end luci.sys.call("rm -f " .. tmp_file) allowInsecure_default = nil filter_keyword_mode_default = uci:get(appname, "@global_subscribe[0]", "filter_keyword_mode") or "0" filter_keyword_discard_list_default = uci:get(appname, "@global_subscribe[0]", "filter_discard_list") or {} filter_keyword_keep_list_default = uci:get(appname, "@global_subscribe[0]", "filter_keep_list") or {} ss_type_default = uci:get(appname, "@global_subscribe[0]", "ss_type") or "shadowsocks-libev" trojan_type_default = uci:get(appname, "@global_subscribe[0]", "trojan_type") or "trojan-plus" vmess_type_default = uci:get(appname, "@global_subscribe[0]", "vmess_type") or "xray" vless_type_default = uci:get(appname, "@global_subscribe[0]", "vless_type") or "xray" hysteria2_type_default = uci:get(appname, "@global_subscribe[0]", "hysteria2_type") or "hysteria2" end if #fail_list > 0 then for index, value in ipairs(fail_list) do log(string.format('【%s】订阅失败,可能是订阅地址无效,或是网络问题,请诊断![%s]', value.remark, tostring(value.http_code))) end end update_node(0) end end if arg[1] then if arg[1] == "start" then log('开始订阅...') xpcall(execute, function(e) log(e) if type(debug) == "table" and type(debug.traceback) == "function" then log(debug.traceback()) end log('发生错误, 正在恢复服务') end) log('订阅完毕...\n') elseif arg[1] == "add" then local f = assert(io.open("/tmp/links.conf", 'r')) local raw = f:read('*all') f:close() parse_link(raw, "1", "导入") update_node(1) luci.sys.call("rm -f /tmp/links.conf") elseif arg[1] == "truncate" then truncate_nodes(arg[2]) end end
294coder/Efficient-MIF
2,476
configs/LEMamba_config.yaml
optimizer: name: "adamw" lr: !!float 1e-3 weight_decay: !!float 1e-6 lr_scheduler: name: 'cos_anneal_restart_reduce' T_0: 50 T_mult: 2 lr_mult: 0.9 eta_min: !!float 8e-5 max_norm: 0.03 ema_decay: 0.995 path: wv3_train_path: <your_path>/train_data/ wv3_val_path: <your_path>/val_data/ gf2_train_path: <your_path>/train_data/ gf2_val_path: <your_path>/val_data/ cave_x8_train_path: <your_path>/train_data/ cave_x8_val_path: <your_path>/val_data/ harvard_x8_train_path: <your_path>/train_data/ harvard_x8_val_path: <your_path>/val_data/ network_configs: LEMamba: img_channel: 4 condition_channel: 1 out_channel: 4 # v2 # width: 32 naf_enc_blk_nums: [] naf_dec_blk_nums: [] naf_chan_upscale: [] middle_blk_nums: 1 # variant tiny (wv3) width: 32 ssm_enc_blk_nums: [2, 1, 1] ssm_dec_blk_nums: [2, 1, 1] ssm_chan_upscale: [1, 1, 1] ssm_ratios: [2, 2, 1] window_sizes: [8,8,null] ssm_enc_d_states: [[16, 32], [16, 32], [null, 32]] ssm_dec_d_states: [[null, 32], [16, 32], [16, 32]] ssm_enc_convs: [[7, 11], [7, 11], [null, 11]] ssm_dec_convs: [[null, 11], [7, 11], [7, 11]] drop_path_rate: 0.1 # small # ssm_enc_blk_nums: [2,2,2] # ssm_dec_blk_nums: [2,2,2] # ssm_chan_upscale: [1,2,2] # ssm_ratios: [3,2,2] # window_sizes: [8,8,null] # ssm_enc_d_states: [[32, 32], [32, 32], [null, 32]] # ssm_dec_d_states: [[null, 32], [32, 32], [32, 32]] # ssm_enc_convs: [[7, 11], [7, 11], [null, 11]] # ssm_dec_convs: [[null, 11], [7, 11], [null, 11]] # drop_path_rate: 0.2 # base # ssm_enc_blk_nums: [4, 3, 2] # ssm_dec_blk_nums: [4, 3, 2] # ssm_chan_upscale: [1, 1, 1] # ssm_ratios: [2,2,1] # window_sizes: [8,8,null] # ssm_enc_d_states: [[16, 32], [16, 32], [null, 32]] # ssm_dec_d_states: [[null, 32], [16, 32], [16, 32]] # ssm_enc_convs: [[7, 11], [7, 11], [null, 11]] # ssm_dec_convs: [[null, 11], [7, 11], [7, 11]] # drop_path_rate: 0.2 if_abs_pos: no if_rope: no patch_merge: yes upscale: 4 LEMamba_only_NAF: img_channel: 8 condition_channel: 1 out_channel: 8 width: 32 naf_enc_blk_nums: [2, 2, 2] naf_dec_blk_nums: [2, 2, 2] naf_chan_upscale: [2, 2, 2] middle_blk_nums: 2 if_abs_pos: no if_rope: no patch_merge: yes upscale: 4 logger_config: base_path: ./log_file/ name: LEMamba file_mode: w
294coder/Efficient-MIF
1,569
configs/dct_config.yaml
optimizer: name: "adamw" lr: !!float 1e-3 weight_decay: !!float 1e-6 lr_scheduler: name: "multi_step" milestones: [200, 1600] gamma: 0.1 # name: 'identity' # name: "plateau" # mode: "min" # threshold: !!float 5e-6 # l1 # min_lr: !!float 1e-6 # patience: 30 # factor: !!float 0.2 # cooldown: 10 max_norm: 0.03 ema_decay: 0.999 path: wv3_train_path: "/volsparse1/dataset/PanCollection/training_data/train_wv3_9714.h5" wv3_val_path: "/volsparse1/dataset/PanCollection/test_data/test_wv3_multiExm1.h5" gf2_train_path: "/volsparse1/dataset/PanCollection/training_data/train_gf2_19809.h5" gf2_val_path: "/volsparse1/dataset/PanCollection/test_data/test_gf2_multiExm1.h5" cave_x4_train_path: "/volsparse1/dataset/HISR/cave_x4/train_cave(with_up)x4.h5" cave_x4_val_path: "/volsparse1/dataset/HISR/cave_x4/test_cave(with_up)x4.h5" cave_x8_train_path: "/volsparse1/dataset/HISR/cave_x8/train_cave(with_up)x8_rgb_16.h5" cave_x8_val_path: "/volsparse1/dataset/HISR/cave_x8/test_cave(with_up)x8_rgb.h5" harvard_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5" harvard_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/test_harvard(with_up)x4_rgb.h5" harvard_x8_train_path: "/volsparse1/dataset/HISR/harvard_x8/train_harvard(with_up)x8_rgb.h5" harvard_x8_val_path: "/volsparse1/dataset/HISR/harvard_x8/test_harvard(with_up)x8_rgb.h5" network_configs: n_colors: 31 upscale_factor: 8 logger_config: base_path: ./log_file/ name: reciprocal_transformer file_mode: w
2977094657/DidaAPI
4,417
routers/habits.py
"""习惯管理相关API路由""" from fastapi import APIRouter from fastapi.responses import Response # 不再需要响应模型导入 from services import habit_service, dida_service from utils import app_logger router = APIRouter(prefix="/habits", tags=["习惯管理"]) @router.get("/all", summary="获取所有习惯", description="获取当前用户的所有习惯列表") async def get_all_habits(): """ 获取所有习惯 返回当前用户的所有习惯列表,包括: - 习惯ID、名称、图标、颜色 - 习惯状态、激励语句、总打卡次数 - 创建时间、修改时间、类型、目标值等信息 **注意**: 需要先完成微信登录获取认证会话 """ try: app_logger.info("请求获取所有习惯") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"} # 获取认证信息 current_session = dida_service.current_session auth_token = current_session['auth_token'] csrf_token = current_session['csrf_token'] # 调用习惯服务 result = await habit_service.get_habits(auth_token, csrf_token) if not result: return {"error": "service_error", "message": "获取习惯列表失败,请稍后重试"} # 记录日志 if 'error' in result: app_logger.info(f"习惯获取失败: {result.get('error')}") else: habit_count = len(result) if isinstance(result, list) else 0 app_logger.info(f"习惯获取完成,习惯数: {habit_count}") # 直接返回原始响应 return result except Exception as e: app_logger.error(f"获取习惯时发生未知错误: {e}") return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"} @router.get("/statistics/week/current", summary="获取本周习惯打卡统计", description="获取本周的习惯打卡统计信息") async def get_week_current_statistics(): """ 获取本周习惯打卡统计 返回本周的习惯打卡统计,包括: - 每日打卡情况 - 习惯完成率 - 连续打卡天数等信息 **注意**: 需要先完成微信登录获取认证会话 """ try: app_logger.info("请求获取本周习惯打卡统计") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"} # 获取认证信息 current_session = dida_service.current_session auth_token = current_session['auth_token'] csrf_token = current_session['csrf_token'] # 调用习惯服务 result = await habit_service.get_week_current_statistics(auth_token, csrf_token) if not result: return {"error": "service_error", "message": "获取本周习惯打卡统计失败,请稍后重试"} # 记录日志 if 'error' in result: app_logger.info(f"本周习惯打卡统计获取失败: {result.get('error')}") else: app_logger.info(f"本周习惯打卡统计获取完成") # 直接返回原始响应 return result except Exception as e: app_logger.error(f"获取本周习惯打卡统计时发生未知错误: {e}") return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"} @router.get("/export", summary="导出习惯数据", description="导出习惯数据为Excel文件") async def export_habits(): """ 导出习惯数据(Excel格式) 导出当前用户的所有习惯数据为Excel文件,包括: - 习惯基本信息 - 打卡记录 - 统计数据等 **注意**: 需要先完成微信登录获取认证会话 """ try: app_logger.info("请求导出习惯数据") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"} # 获取认证信息 current_session = dida_service.current_session auth_token = current_session['auth_token'] csrf_token = current_session['csrf_token'] # 调用习惯服务 result = await habit_service.export_habits(auth_token, csrf_token) if not result: return {"error": "service_error", "message": "导出习惯数据失败,请稍后重试"} # 检查是否有错误 if 'error' in result: app_logger.info(f"习惯数据导出失败: {result.get('error')}") return result # 成功获取文件内容 app_logger.info(f"习惯数据导出完成,文件名: {result.get('filename')}") # 返回文件响应 return Response( content=result['content'], media_type=result['content_type'], headers={ "Content-Disposition": f"attachment; filename={result['filename']}" } ) except Exception as e: app_logger.error(f"导出习惯数据时发生未知错误: {e}") return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
2977094657/DidaAPI
1,528
routers/projects.py
"""清单管理相关API路由""" from fastapi import APIRouter from services import project_service, dida_service from utils import app_logger router = APIRouter(prefix="/projects", tags=["清单管理"]) @router.get("/all", summary="获取所有项目/清单", description="获取当前用户的所有项目/清单列表") async def get_all_projects(): """ 获取所有项目/清单 返回当前用户的所有项目/清单列表,包括: - 项目ID、名称、颜色 - 项目权限、类型、用户数量 - 创建时间、修改时间等信息 **注意**: 需要先完成微信登录获取认证会话 """ try: app_logger.info("请求获取所有项目") # 检查认证状态 session_status = dida_service.get_session_status() if not session_status["has_session"]: return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"} # 获取认证信息 current_session = dida_service.current_session auth_token = current_session['auth_token'] csrf_token = current_session['csrf_token'] # 调用项目服务 result = await project_service.get_projects(auth_token, csrf_token) if not result: return {"error": "service_error", "message": "获取项目列表失败,请稍后重试"} # 记录日志 if 'error' in result: app_logger.info(f"项目获取失败: {result.get('error')}") else: project_count = len(result) if isinstance(result, list) else 0 app_logger.info(f"项目获取完成,项目数: {project_count}") # 直接返回原始响应 return result except Exception as e: app_logger.error(f"获取项目时发生未知错误: {e}") return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
294coder/Efficient-MIF
2,223
configs/MIMO_SST_config.yaml
optimizer: name: "adamw" lr: !!float 1e-4 weight_decay: !!float 1e-6 lr_scheduler: # name: 'cos_anneal' # name: 'cos_anneal_restart' # T_0: 50 # T_mult: 2 # eta_min: !!float 1e-9 # T_max: 2000 # eta_min: !!float 1e-8 # name: "plateau" # mode: "min" # threshold: !!float 1e-5 # l1 # min_lr: !!float 1e-6 # patience: 20 # factor: !!float 0.1 # cooldown: 5 # threshold_mode: "rel" name: 'identity' # name: "multi_step" # milestones: [1000] #[100, 800] # gamma: 0.2 max_norm: 0.03 ema_decay: 0.999 path: # pansharpening wv3_train_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5" wv3_val_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5" gf2_train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5" gf2_val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5" qb_train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5" qb_val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5" # hyperspectral multispectral image fusion cave_x4_train_path: "/volsparse1/dataset/HISR/cave_x8/train_cave(with_up)x8_rgb_16.h5" # val_path: "/home/ShangqiDeng/data/HSI/cave_x4/test_cavepatches64-2.h5" cave_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5" cave_x8_train_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/train_cave(with_up)x8_rgb_16.h5" # val_path: "/home/ZiHanCao/datasets/HISI/new_cave/x8/validation_cave(with_up)x8_rgb.h5" cave_x8_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/test_cave(with_up)x8_rgb.h5" harvard_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5" harvard_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/test_harvard(with_up)x4_rgb.h5" harvard_x8_train_path: "/volsparse1/dataset/HISR/harvard_x8/train_harvard(with_up)x8_rgb.h5" harvard_x8_val_path: "/volsparse1/dataset/HISR/harvard_x8/test_harvard(with_up)x8_rgb.h5" network_configs: in_chan: 8 logger_config: base_path: ./log_file/ name: MIM_SST file_mode: w
294coder/Efficient-MIF
3,581
configs/pmacnet_config.yaml
optimizer: name: "adamw" lr: !!float 3e-4 weight_decay: !!float 1e-6 lr_scheduler: # name: 'multi_step' # milestones: [400, 800] # gamma: 0.1 name: "identity" max_norm: null path: # train_path: '/Data2/DataSet/HISR/CAVE/train_cave(with_up)x4.h5' # val_path: '/Data2/DataSet/HISR/CAVE/validation_cave(with_up)x4.h5' # train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5' # val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5' # train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5' # val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/training_gf2/valid_gf2.h5' # train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5' # val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/training_qb/valid_qb.h5' # train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5" # val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/valid_gf2.h5" train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5" val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5" # train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5' # val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5' # train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5' # val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5' # train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5' # val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5' # train_path: '/home/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5' # val_path: '/home/ZiHanCao/datasets/pansharpening/gf/training_gf2/valid_gf2.h5' # train_path: '/home/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5' # val_path: '/home/ZiHanCao/datasets/pansharpening/qb/training_qb/valid_qb.h5' # train_path: "/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/HISI/new_cave/train_cave(with_up)x4.h5" # val_path: "/media/office-401-remote/Elements SE/cao/cavex4/test_cavepatches64-2.h5" # train_path: "/home/ShangqiDeng/data/HSI/cave_x8/train_cave(with_up)x8_rgb.h5" # val_path: "/home/ZiHanCao/datasets/HISI/new_cave/x8/validation_cave(with_up)x8_rgb.h5" # train_path: "/home/ShangqiDeng/data/HSI/cave_x8/train_cave(with_up)x8_rgb_16.h5" # val_path: "/home/ShangqiDeng/data/HSI/cave_x8/test_cavepatches128-2.h5" # train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/train_harvard(with_up)x8_rgb.h5" # val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/test_clip_128.h5" # train_path: '/home/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5' # val_path: '/home/ZiHanCao/datasets/HISI/new_harvard/validation_harvard(with_up)x4_rgb.h5' network_configs: ms_inp_ch: 4 num_layers: 4 latent_dim: 64 logger_config: base_path: ./log_file/ name: pmacnet file_mode: w
2977094657/DidaAPI
7,590
routers/auth.py
"""认证相关API路由""" from fastapi import APIRouter, HTTPException, Query from fastapi.responses import FileResponse from models import WeChatQRResponse, WeChatValidateResponse, ApiResponse, PasswordLoginRequest from services import wechat_service from utils import app_logger import os router = APIRouter(prefix="/auth", tags=["认证"]) @router.get("/wechat/login", summary="微信扫码登录页面", description="返回完整的微信扫码登录HTML页面") async def wechat_login_page(): """ 微信扫码登录页面 返回一个完整的HTML页面,集成了: - 自动获取微信二维码 - 实时显示二维码 - 自动轮询登录状态 - 登录成功处理 - 错误处理和重试 用户只需访问此页面即可完成完整的微信扫码登录流程 """ try: static_file = os.path.join("static", "wechat_login.html") if not os.path.exists(static_file): raise HTTPException( status_code=404, detail="微信登录页面文件不存在" ) return FileResponse( static_file, media_type="text/html", headers={"Cache-Control": "no-cache"} ) except Exception as e: app_logger.error(f"返回微信登录页面时发生错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/wechat/qrcode", response_model=WeChatQRResponse, summary="获取微信登录二维码", description="获取微信登录二维码,用户扫码后可进行登录验证") async def get_wechat_qrcode( state: str = Query(default="Lw==", description="状态参数,用于防止CSRF攻击") ) -> WeChatQRResponse: """ 获取微信登录二维码 - **state**: 状态参数,默认为 "Lw==",用于防止CSRF攻击 返回包含二维码图片URL和密钥的响应 """ try: app_logger.info(f"请求获取微信二维码,state: {state}") qr_response = await wechat_service.get_qr_code(state) if not qr_response: raise HTTPException( status_code=500, detail="获取微信二维码失败,请稍后重试" ) app_logger.info(f"成功返回微信二维码: {qr_response.qr_code_key}") return qr_response except HTTPException: raise except Exception as e: app_logger.error(f"获取微信二维码时发生未知错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/wechat/validate", response_model=WeChatValidateResponse, summary="验证微信登录", description="使用扫码后获得的验证码进行微信登录验证") async def validate_wechat_login( code: str = Query(..., description="扫码后获得的验证码"), state: str = Query(default="Lw==", description="状态参数,需与获取二维码时的参数一致") ) -> WeChatValidateResponse: """ 验证微信登录 - **code**: 扫码后获得的验证码(必需) - **state**: 状态参数,需与获取二维码时的参数一致 返回登录验证结果,包含用户信息和认证令牌 """ try: app_logger.info(f"请求验证微信登录,code: {code}, state: {state}") if not code: raise HTTPException( status_code=400, detail="验证码不能为空" ) validate_response = await wechat_service.validate_wechat_login(code, state) if not validate_response: raise HTTPException( status_code=500, detail="验证微信登录失败,请稍后重试" ) app_logger.info(f"微信登录验证完成,成功: {validate_response.success}") return validate_response except HTTPException: raise except Exception as e: app_logger.error(f"验证微信登录时发生未知错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/wechat/poll", summary="轮询微信登录状态", description="轮询检查二维码是否已被扫码登录") async def poll_wechat_login( qr_code_key: str = Query(..., description="二维码密钥(16位字符)"), max_attempts: int = Query(default=60, description="最大轮询次数,默认60次(约5分钟)") ) -> WeChatValidateResponse: """ 轮询微信登录状态 - **qr_code_key**: 二维码密钥(16位字符) - **max_attempts**: 最大轮询次数,默认60次(约5分钟) 这个接口会持续轮询微信服务器,检查二维码是否已被扫码登录 """ try: app_logger.info(f"开始轮询微信登录状态,qr_code_key: {qr_code_key}") result = await wechat_service.poll_qr_status(qr_code_key, max_attempts) if not result: raise HTTPException( status_code=500, detail="轮询微信登录状态失败" ) app_logger.info(f"轮询完成,结果: {result.success}") return result except HTTPException: raise except Exception as e: app_logger.error(f"轮询微信登录状态时发生未知错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/wechat/callback", summary="微信登录回调处理", description="处理微信扫码后的回调,提取code参数") async def wechat_callback( code: str = Query(..., description="微信返回的授权码"), state: str = Query(default="Lw==", description="状态参数") ): """ 微信登录回调处理 这个接口用于演示微信回调的处理流程。 实际使用中,微信会重定向到配置的redirect_uri。 - **code**: 微信返回的授权码 - **state**: 状态参数 """ try: app_logger.info(f"收到微信回调,code: {code}, state: {state}") # 自动进行登录验证 validate_response = await wechat_service.validate_wechat_login(code, state) if validate_response and validate_response.success: return { "message": "微信登录成功", "code": code, "state": state, "login_result": validate_response.dict(), "next_step": "用户已成功登录,可以访问受保护的资源" } else: return { "message": "微信登录失败", "code": code, "state": state, "error": validate_response.message if validate_response else "验证失败" } except Exception as e: app_logger.error(f"处理微信回调时发生错误: {e}") return { "message": "处理微信回调失败", "code": code, "state": state, "error": str(e) } @router.post("/password/login", summary="密码登录", description="使用手机号/邮箱,密码进行登录") async def password_login( login_request: PasswordLoginRequest ): """ 密码登录滴答清单 - **username**: 登录账户(邮箱或手机号) - **password**: 登录密码 返回登录结果,包含用户信息和认证令牌 """ try: app_logger.info(f"请求密码登录,用户名: {login_request.username}") if not login_request.username or not login_request.password: raise HTTPException( status_code=400, detail="用户名和密码不能为空" ) login_response = await wechat_service.password_login( login_request.username, login_request.password ) if not login_response: return {"error": "密码登录失败,请稍后重试"} # 记录日志 if 'errorCode' in login_response: app_logger.info(f"密码登录失败,错误代码: {login_response.get('errorCode')}") else: user_id = login_response.get('userId', '') app_logger.info(f"密码登录成功,用户ID: {user_id}") # 直接返回原始响应,不进行任何封装 return login_response except HTTPException: raise except Exception as e: app_logger.error(f"密码登录时发生未知错误: {e}") raise HTTPException( status_code=500, detail=f"服务器内部错误: {str(e)}" ) @router.get("/health", response_model=ApiResponse, summary="健康检查", description="检查认证服务的健康状态") async def health_check() -> ApiResponse: """ 健康检查接口 返回服务的健康状态信息 """ return ApiResponse( code=200, message="认证服务运行正常", data={ "service": "auth", "status": "healthy", "version": "1.0.0" } )
2977094657/DidaAPI
1,193
utils/logger.py
"""日志配置模块""" import sys import os from datetime import datetime from loguru import logger def setup_logger(): """配置日志系统""" # 移除默认的日志处理器 logger.remove() # 使用默认配置 level = 'DEBUG' format_str = '{time:YYYY-MM-DD HH:mm:ss} | {level} | {name}:{function}:{line} - {message}' rotation = '1 day' retention = '7 days' # 控制台输出 logger.add( sys.stdout, format=format_str, level=level, colorize=True ) # 获取当前日期,用于创建日志文件夹结构 now = datetime.now() year = now.strftime('%Y') month = now.strftime('%m') day = now.strftime('%d') # 创建日志文件夹结构: output/logs/年/月/日/ log_dir = f"output/logs/{year}/{month}/{day}" os.makedirs(log_dir, exist_ok=True) # 应用日志文件 logger.add( f"{log_dir}/app.log", format=format_str, level=level, rotation=rotation, retention=retention, encoding="utf-8" ) # 错误日志单独文件 logger.add( f"{log_dir}/error.log", format=format_str, level="ERROR", rotation=rotation, retention=retention, encoding="utf-8" ) return logger # 初始化日志 app_logger = setup_logger()
281677160/openwrt-package
17,188
luci-app-passwall/root/usr/share/passwall/rule_update.lua
#!/usr/bin/lua local api = require ("luci.passwall.api") local name = api.appname local uci = api.uci local sys = api.sys local jsonc = api.jsonc local fs = api.fs local arg1 = arg[1] local arg2 = arg[2] local arg3 = arg[3] local nftable_name = "inet passwall" local rule_path = "/usr/share/" .. name .. "/rules" local reboot = 0 local gfwlist_update = "0" local chnroute_update = "0" local chnroute6_update = "0" local chnlist_update = "0" local geoip_update = "0" local geosite_update = "0" -- match comments/title/whitelist/ip address/excluded_domain local comment_pattern = "^[#!\\[@]+" local ip_pattern = "^%d+%.%d+%.%d+%.%d+" local ip4_ipset_pattern = "^%d+%.%d+%.%d+%.%d+[%/][%d]+$" local ip6_ipset_pattern = ":-[%x]+%:+[%x]-[%/][%d]+$" local domain_pattern = "([%w%-]+%.[%w%.%-]+)[%/%*]*" local excluded_domain = {"apple.com","sina.cn","sina.com.cn","baidu.com","byr.cn","jlike.com","weibo.com","zhongsou.com","youdao.com","sogou.com","so.com","soso.com","aliyun.com","taobao.com","jd.com","qq.com","bing.com"} local gfwlist_url = uci:get(name, "@global_rules[0]", "gfwlist_url") or {"https://fastly.jsdelivr.net/gh/Loyalsoldier/v2ray-rules-dat@release/gfw.txt"} local chnroute_url = uci:get(name, "@global_rules[0]", "chnroute_url") or {"https://ispip.clang.cn/all_cn.txt"} local chnroute6_url = uci:get(name, "@global_rules[0]", "chnroute6_url") or {"https://ispip.clang.cn/all_cn_ipv6.txt"} local chnlist_url = uci:get(name, "@global_rules[0]", "chnlist_url") or {"https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/accelerated-domains.china.conf","https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/apple.china.conf","https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/google.china.conf"} local geoip_url = uci:get(name, "@global_rules[0]", "geoip_url") or "https://github.com/Loyalsoldier/geoip/releases/latest/download/geoip.dat" local geosite_url = uci:get(name, "@global_rules[0]", "geosite_url") or "https://github.com/Loyalsoldier/v2ray-rules-dat/releases/latest/download/geosite.dat" local asset_location = uci:get(name, "@global_rules[0]", "v2ray_location_asset") or "/usr/share/v2ray/" local use_nft = uci:get(name, "@global_forwarding[0]", "use_nft") or "0" local geo2rule = uci:get(name, "@global_rules[0]", "geo2rule") or "0" local geoip_update_ok, geosite_update_ok = false, false asset_location = asset_location:match("/$") and asset_location or (asset_location .. "/") --兼容旧版本geo下载方式的配置,择机删除。 if geoip_url:match(".*/([^/]+)$") == "latest" then geoip_url = "https://github.com/Loyalsoldier/geoip/releases/latest/download/geoip.dat" end if geosite_url:match(".*/([^/]+)$") == "latest" then geosite_url = "https://github.com/Loyalsoldier/v2ray-rules-dat/releases/latest/download/geosite.dat" end if arg3 == "cron" then arg2 = nil end local log = function(...) if arg1 then if arg1 == "log" then api.log(...) elseif arg1 == "print" then local result = os.date("%Y-%m-%d %H:%M:%S: ") .. table.concat({...}, " ") print(result) end end end local function gen_nftset(set_name, ip_type, tmp_file, input_file) f = io.open(input_file, "r") local element = f:read("*all") f:close() nft_file, err = io.open(tmp_file, "w") nft_file:write('#!/usr/sbin/nft -f\n') nft_file:write(string.format('define %s = {%s}\n', set_name, string.gsub(element, "%s*%c+", " timeout 3650d, "))) if sys.call(string.format('nft "list set %s %s" >/dev/null 2>&1', nftable_name, set_name)) ~= 0 then nft_file:write(string.format('add set %s %s { type %s; flags interval, timeout; timeout 2d; gc-interval 2d; auto-merge; }\n', nftable_name, set_name, ip_type)) end nft_file:write(string.format('add element %s %s $%s\n', nftable_name, set_name, set_name)) nft_file:close() sys.call(string.format('nft -f %s &>/dev/null',tmp_file)) os.remove(tmp_file) end --gen cache for nftset from file local function gen_cache(set_name, ip_type, input_file, output_file) local tmp_dir = "/tmp/" local tmp_file = output_file .. "_tmp" local tmp_set_name = set_name .. "_tmp" gen_nftset(tmp_set_name, ip_type, tmp_file, input_file) sys.call(string.format('nft list set %s %s | sed "s/%s/%s/g" | cat > %s', nftable_name, tmp_set_name, tmp_set_name, set_name, output_file)) sys.call(string.format('nft flush set %s %s', nftable_name, tmp_set_name)) sys.call(string.format('nft delete set %s %s', nftable_name, tmp_set_name)) end -- curl local function curl(url, file, valifile) local args = { "-skL", "-w %{http_code}", "--retry 3", "--connect-timeout 3", "--max-time 300", "--speed-limit 51200 --speed-time 15" } if file then args[#args + 1] = "-o " .. file end if valifile then args[#args + 1] = "--dump-header " .. valifile end local return_code, result = api.curl_auto(url, nil, args) return tonumber(result) end --check excluded domain local function check_excluded_domain(value) for k,v in ipairs(excluded_domain) do if value:find(v) then return true end end end local function line_count(file_path) local num = 0 for _ in io.lines(file_path) do num = num + 1 end return num; end local function non_file_check(file_path, vali_file) if fs.readfile(file_path, 10) then local size_str = sys.exec("grep -i 'Content-Length' " .. vali_file .. " | tail -n1 | sed 's/[^0-9]//g'") local remote_file_size = tonumber(size_str ~= "" and size_str or nil) local local_file_size = tonumber(fs.stat(file_path, "size")) if remote_file_size and local_file_size then if remote_file_size == local_file_size then return nil; else log("下载文件大小校验出错,原始文件大小" .. remote_file_size .. "B,下载文件大小:" .. local_file_size .. "B。") return true; end else return nil; end else log("下载文件读取出错。") return true; end end local function GeoToRule(rule_name, rule_type, out_path) if not api.is_finded("geoview") then log(rule_name .. "生成失败,缺少 geoview 组件。") return false; end local geosite_path = asset_location .. "geosite.dat" local geoip_path = asset_location .. "geoip.dat" local file_path = (rule_type == "domain") and geosite_path or geoip_path local arg if rule_type == "domain" then if rule_name == "gfwlist" then arg = "-type geosite -list gfw" else arg = "-type geosite -list cn" end elseif rule_type == "ip4" then arg = "-type geoip -list cn -ipv6=false" elseif rule_type == "ip6" then arg = "-type geoip -list cn -ipv4=false" end cmd = string.format("geoview -input '%s' %s -lowmem=true -output '%s'", file_path, arg, out_path) sys.exec(cmd) return true; end --fetch rule local function fetch_rule(rule_name,rule_type,url,exclude_domain) local sret = 200 local sret_tmp = 0 local domains = {} local file_tmp = "/tmp/" ..rule_name.. "_tmp" local vali_file = "/tmp/" ..rule_name.. "_vali" local download_file_tmp = "/tmp/" ..rule_name.. "_dl" local unsort_file_tmp = "/tmp/" ..rule_name.. "_unsort" if geo2rule == "1" then url = {"geo2rule"} log(rule_name.. " 开始生成...") else log(rule_name.. " 开始更新...") end for k,v in ipairs(url) do if v ~= "geo2rule" then sret_tmp = curl(v, download_file_tmp..k, vali_file..k) if sret_tmp == 200 and non_file_check(download_file_tmp..k, vali_file..k) then log(rule_name.. " 第" ..k.. "条规则:" ..v.. "下载文件过程出错,尝试重新下载。") os.remove(download_file_tmp..k) os.remove(vali_file..k) sret_tmp = curl(v, download_file_tmp..k, vali_file..k) if sret_tmp == 200 and non_file_check(download_file_tmp..k, vali_file..k) then sret = 0 sret_tmp = 0 log(rule_name.. " 第" ..k.. "条规则:" ..v.. "下载文件过程出错,请检查网络或下载链接后重试!") end end else if not GeoToRule(rule_name, rule_type, download_file_tmp..k) then return 1 end sret_tmp = 200 end if sret_tmp == 200 then if rule_name == "gfwlist" and geo2rule == "0" then local domains = {} local gfwlist = io.open(download_file_tmp..k, "r") local decode = api.base64Decode(gfwlist:read("*all")) gfwlist:close() gfwlist = io.open(download_file_tmp..k, "w") gfwlist:write(decode) gfwlist:close() end if rule_type == "domain" and exclude_domain == true then for line in io.lines(download_file_tmp..k) do line = line:gsub("full:", "") if not (string.find(line, comment_pattern) or string.find(line, ip_pattern) or check_excluded_domain(line) or string.find(line, ":")) then local match = string.match(line, domain_pattern) if match then domains[match] = true end end end elseif rule_type == "domain" then for line in io.lines(download_file_tmp..k) do line = line:gsub("full:", "") if not (string.find(line, comment_pattern) or string.find(line, ip_pattern) or string.find(line, ":")) then local match = string.match(line, domain_pattern) if match then domains[match] = true end end end elseif rule_type == "ip4" then local out = io.open(unsort_file_tmp, "a") for line in io.lines(download_file_tmp..k) do if string.match(line, ip4_ipset_pattern) and not string.match(line, "^0%..*") then out:write(string.format("%s\n", line)) end end out:close() elseif rule_type == "ip6" then local out = io.open(unsort_file_tmp, "a") for line in io.lines(download_file_tmp..k) do if string.match(line, ip6_ipset_pattern) and not string.match(line, "^::(/%d+)?$") then out:write(string.format("%s\n", line)) end end out:close() end else sret = 0 log(rule_name.. " 第" ..k.. "条规则:" ..v.. "下载失败,请检查网络或下载链接后重试!") end os.remove(download_file_tmp..k) os.remove(vali_file..k) end if sret == 200 then if rule_type == "domain" then local out = io.open(unsort_file_tmp, "w") for k,v in pairs(domains) do out:write(string.format("%s\n", k)) end out:close() end sys.call("LC_ALL=C sort -u " .. unsort_file_tmp .. " > " .. file_tmp) os.remove(unsort_file_tmp) local old_md5 = sys.exec("echo -n $(md5sum " .. rule_path .. "/" ..rule_name.. " | awk '{print $1}')"):gsub("\n", "") local new_md5 = sys.exec("echo -n $([ -f '" ..file_tmp.. "' ] && md5sum " ..file_tmp.." | awk '{print $1}')"):gsub("\n", "") if old_md5 ~= new_md5 then local count = line_count(file_tmp) if use_nft == "1" and (rule_type == "ip6" or rule_type == "ip4") then local output_file = file_tmp.. ".nft" if rule_type == "ip4" then local set_name = "passwall_" ..rule_name if rule_name == "chnroute" then set_name = "passwall_chn" end gen_cache(set_name, "ipv4_addr", file_tmp, output_file) elseif rule_type == "ip6" then local set_name = "passwall_" ..rule_name if rule_name == "chnroute6" then set_name = "passwall_chn6" end gen_cache(set_name, "ipv6_addr", file_tmp, output_file) end sys.exec(string.format('mv -f %s %s', output_file, rule_path .. "/" ..rule_name.. ".nft")) os.remove(output_file) end sys.exec("mv -f "..file_tmp .. " " ..rule_path .. "/" ..rule_name) reboot = 1 log(rule_name.. " 更新成功,总规则数 " ..count.. " 条。") else log(rule_name.. " 版本一致,无需更新。") end else log(rule_name.. " 文件下载失败!") end os.remove(file_tmp) return 0 end local function fetch_geofile(geo_name, geo_type, url) local tmp_path = "/tmp/" .. geo_name local asset_path = asset_location .. geo_name local down_filename = url:match("^.*/([^/?#]+)") local sha_url = url:gsub(down_filename, down_filename .. ".sha256sum") local sha_path = tmp_path .. ".sha256sum" local vali_file = tmp_path .. ".vali" local function verify_sha256(sha_file) return sys.call("sha256sum -c " .. sha_file .. " > /dev/null 2>&1") == 0 end local sha_verify = curl(sha_url, sha_path) == 200 if sha_verify then local f = io.open(sha_path, "r") if f then local content = f:read("*l") f:close() if content then content = content:gsub(down_filename, tmp_path) f = io.open(sha_path, "w") if f then f:write(content) f:close() end end end if fs.access(asset_path) then sys.call(string.format("cp -f %s %s", asset_path, tmp_path)) if verify_sha256(sha_path) then log(geo_type .. " 版本一致,无需更新。") return 0 end end end local sret_tmp = curl(url, tmp_path, vali_file) if sret_tmp == 200 and non_file_check(tmp_path, vali_file) then log(geo_type .. " 下载文件过程出错,尝试重新下载。") os.remove(tmp_path) os.remove(vali_file) sret_tmp = curl(url, tmp_path, vali_file) if sret_tmp == 200 and non_file_check(tmp_path, vali_file) then sret_tmp = 0 log(geo_type .. " 下载文件过程出错,请检查网络或下载链接后重试!") end end if sret_tmp == 200 then if sha_verify then if verify_sha256(sha_path) then sys.call(string.format("mkdir -p %s && cp -f %s %s", asset_location, tmp_path, asset_path)) reboot = 1 log(geo_type .. " 更新成功。") if geo_type == "geoip" then geoip_update_ok = true else geosite_update_ok = true end else log(geo_type .. " 更新失败,请稍后重试或更换更新URL。") return 1 end else if fs.access(asset_path) and sys.call(string.format("cmp -s %s %s", tmp_path, asset_path)) == 0 then log(geo_type .. " 版本一致,无需更新。") return 0 end sys.call(string.format("mkdir -p %s && cp -f %s %s", asset_location, tmp_path, asset_path)) reboot = 1 log(geo_type .. " 更新成功。") if geo_type == "geoip" then geoip_update_ok = true else geosite_update_ok = true end end else log(geo_type .. " 更新失败,请稍后重试或更换更新URL。") return 1 end return 0 end local function fetch_gfwlist() fetch_rule("gfwlist","domain",gfwlist_url,true) end local function fetch_chnroute() fetch_rule("chnroute","ip4",chnroute_url,false) end local function fetch_chnroute6() fetch_rule("chnroute6","ip6",chnroute6_url,false) end local function fetch_chnlist() fetch_rule("chnlist","domain",chnlist_url,false) end local function fetch_geoip() fetch_geofile("geoip.dat","geoip",geoip_url) end local function fetch_geosite() fetch_geofile("geosite.dat","geosite",geosite_url) end if arg2 then string.gsub(arg2, '[^' .. "," .. ']+', function(w) if w == "gfwlist" then gfwlist_update = "1" end if w == "chnroute" then chnroute_update = "1" end if w == "chnroute6" then chnroute6_update = "1" end if w == "chnlist" then chnlist_update = "1" end if w == "geoip" then geoip_update = "1" end if w == "geosite" then geosite_update = "1" end end) else gfwlist_update = uci:get(name, "@global_rules[0]", "gfwlist_update") or "1" chnroute_update = uci:get(name, "@global_rules[0]", "chnroute_update") or "1" chnroute6_update = uci:get(name, "@global_rules[0]", "chnroute6_update") or "1" chnlist_update = uci:get(name, "@global_rules[0]", "chnlist_update") or "1" geoip_update = uci:get(name, "@global_rules[0]", "geoip_update") or "1" geosite_update = uci:get(name, "@global_rules[0]", "geosite_update") or "1" end if gfwlist_update == "0" and chnroute_update == "0" and chnroute6_update == "0" and chnlist_update == "0" and geoip_update == "0" and geosite_update == "0" then os.exit(0) end log("开始更新规则...") local function safe_call(func, err_msg) xpcall(func, function(e) log(e) log(debug.traceback()) log(err_msg) end) end local function remove_tmp_geofile(name) os.remove("/tmp/" .. name .. ".dat") os.remove("/tmp/" .. name .. ".dat.sha256sum") os.remove("/tmp/" .. name .. ".dat.vali") end if geo2rule == "1" then if geoip_update == "1" then log("geoip 开始更新...") safe_call(fetch_geoip, "更新geoip发生错误...") remove_tmp_geofile("geoip") end if geosite_update == "1" then log("geosite 开始更新...") safe_call(fetch_geosite, "更新geosite发生错误...") remove_tmp_geofile("geosite") end if geoip_update_ok then safe_call(fetch_chnroute, "生成chnroute发生错误...") safe_call(fetch_chnroute6, "生成chnroute6发生错误...") end if geosite_update_ok then safe_call(fetch_gfwlist, "生成gfwlist发生错误...") safe_call(fetch_chnlist, "生成chnlist发生错误...") end else if gfwlist_update == "1" then safe_call(fetch_gfwlist, "更新gfwlist发生错误...") end if chnroute_update == "1" then safe_call(fetch_chnroute, "更新chnroute发生错误...") end if chnroute6_update == "1" then safe_call(fetch_chnroute6, "更新chnroute6发生错误...") end if chnlist_update == "1" then safe_call(fetch_chnlist, "更新chnlist发生错误...") end if geoip_update == "1" then log("geoip 开始更新...") safe_call(fetch_geoip, "更新geoip发生错误...") remove_tmp_geofile("geoip") end if geosite_update == "1" then log("geosite 开始更新...") safe_call(fetch_geosite, "更新geosite发生错误...") remove_tmp_geofile("geosite") end end uci:set(name, "@global_rules[0]", "gfwlist_update", gfwlist_update) uci:set(name, "@global_rules[0]", "chnroute_update", chnroute_update) uci:set(name, "@global_rules[0]", "chnroute6_update", chnroute6_update) uci:set(name, "@global_rules[0]", "chnlist_update", chnlist_update) uci:set(name, "@global_rules[0]", "geoip_update", geoip_update) uci:set(name, "@global_rules[0]", "geosite_update", geosite_update) api.uci_save(uci, name, true) if reboot == 1 then if arg3 == "cron" then if not fs.access("/var/lock/" .. name .. ".lock") then sys.call("touch /tmp/lock/" .. name .. "_cron.lock") end end log("重启服务,应用新的规则。") uci:set(name, "@global[0]", "flush_set", "1") api.uci_save(uci, name, true, true) end log("规则更新完毕...\n")
2977094657/DidaAPI
2,107
services/project_service.py
"""项目管理服务模块""" import httpx from typing import Optional from utils import app_logger from core import urls # 不再使用响应模型,直接返回原始响应 class ProjectService: """项目管理服务类""" def __init__(self): self.client = httpx.AsyncClient(timeout=30.0) async def get_projects(self, auth_token: str, csrf_token: str) -> dict: """ 获取项目/清单列表 Args: auth_token: 认证令牌 csrf_token: CSRF令牌 Returns: dict: 原始响应数据 """ try: # 构建请求URL url = urls.build_dida_api_url(urls.DIDA_PROJECT_APIS["get_projects"]) # 构建请求头 headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36', 'Accept': 'application/json, text/plain, */*', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8', 'Content-Type': 'application/json', 'X-Requested-With': 'XMLHttpRequest', 'X-Tz': 'Asia/Shanghai', } # 构建cookies cookies = { 't': auth_token, '_csrf_token': csrf_token } app_logger.info(f"请求获取项目列表: {url}") # 发送请求 response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: response_data = response.json() app_logger.info(f"成功获取项目列表,项目数: {len(response_data) if isinstance(response_data, list) else 0}") # 直接返回原始响应 return response_data else: app_logger.error(f"获取项目列表失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"获取项目列表时发生错误: {e}") return {"error": str(e)} async def close(self): """关闭HTTP客户端""" await self.client.aclose() # 全局项目服务实例 project_service = ProjectService()
2977094657/DidaAPI
6,838
services/pomodoro_service.py
"""番茄专注服务模块""" import httpx from datetime import datetime, timezone, timedelta from core import urls class PomodoroService: """番茄专注服务类""" def __init__(self): self.client = httpx.AsyncClient(timeout=30.0) def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict: """构建认证请求头""" return { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36', 'Accept': 'application/json, text/plain, */*', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8', 'Content-Type': 'application/json', 'X-Requested-With': 'XMLHttpRequest', 'X-Tz': 'Asia/Shanghai', } def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict: """构建认证cookies""" return { 't': auth_token, '_csrf_token': csrf_token } def _convert_time_to_timestamp(self, time_str: str) -> int: """ 将时间字符串转换为时间戳(毫秒) Args: time_str: 时间字符串,格式如 "2025-04-22T08:43:31.000+0000" Returns: int: 毫秒时间戳 """ try: # 解析时间字符串 dt = datetime.fromisoformat(time_str.replace('Z', '+00:00')) # 转换为中国时间(UTC+8) china_tz = timezone(timedelta(hours=8)) china_time = dt.astimezone(china_tz) # 转换为时间戳(秒)然后转为毫秒 timestamp_ms = int(china_time.timestamp() * 1000) return timestamp_ms except Exception as e: raise ValueError(f"时间转换失败: {e}") async def get_general_for_desktop(self, auth_token: str, csrf_token: str) -> dict: """获取番茄专注概览(桌面版),直接返回原始响应""" try: url = urls.build_dida_api_url(urls.DIDA_POMODORO_APIS["general_for_desktop"]) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def get_focus_distribution(self, auth_token: str, csrf_token: str, start_date: str, end_date: str) -> dict: """获取专注详情分布,直接返回原始响应""" try: endpoint = f"{urls.DIDA_POMODORO_APIS['focus_distribution']}/{start_date}/{end_date}" url = urls.build_dida_api_url(endpoint) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def get_focus_timeline(self, auth_token: str, csrf_token: str, to_timestamp: int = None) -> dict: """ 获取专注记录时间线,支持分页 Args: auth_token: 认证令牌 csrf_token: CSRF令牌 to_timestamp: 可选的时间戳参数,用于分页获取更早的数据 Returns: dict: 原始API响应 """ try: url = urls.build_dida_api_url(urls.DIDA_POMODORO_APIS["focus_timeline"]) # 如果提供了时间戳参数,添加到URL中 if to_timestamp is not None: url = f"{url}?to={to_timestamp}" headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def get_focus_heatmap(self, auth_token: str, csrf_token: str, start_date: str, end_date: str) -> dict: """获取专注趋势热力图,直接返回原始响应""" try: endpoint = f"{urls.DIDA_POMODORO_APIS['focus_heatmap']}/{start_date}/{end_date}" url = urls.build_dida_api_url(endpoint) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def get_focus_time_distribution(self, auth_token: str, csrf_token: str, start_date: str, end_date: str) -> dict: """获取专注时间分布(按时间段),直接返回原始响应""" try: endpoint = f"{urls.DIDA_POMODORO_APIS['focus_time_distribution']}/{start_date}/{end_date}" url = urls.build_dida_api_url(endpoint) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def get_focus_hour_distribution(self, auth_token: str, csrf_token: str, start_date: str, end_date: str) -> dict: """获取专注时间按小时分布,直接返回原始响应""" try: endpoint = f"{urls.DIDA_POMODORO_APIS['focus_hour_distribution']}/{start_date}/{end_date}" url = urls.build_dida_api_url(endpoint) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def close(self): """关闭HTTP客户端""" await self.client.aclose() # 全局番茄专注服务实例 pomodoro_service = PomodoroService()
281677160/openwrt-package
3,284
luci-app-passwall/root/usr/share/passwall/test.sh
#!/bin/sh CONFIG=passwall LOG_FILE=/tmp/log/$CONFIG.log echolog() { local d="$(date "+%Y-%m-%d %H:%M:%S")" #echo -e "$d: $1" echo -e "$d: $1" >> $LOG_FILE } config_n_get() { local ret=$(uci -q get "${CONFIG}.${1}.${2}" 2>/dev/null) echo "${ret:=$3}" } lua_api() { local func=${1} [ -z "${func}" ] && { echo "nil" return } echo $(lua -e "local api = require 'luci.passwall.api' print(api.${func})") } test_url() { local url=$1 local try=1 [ -n "$2" ] && try=$2 local timeout=2 [ -n "$3" ] && timeout=$3 local extra_params=$4 curl --help all | grep "\-\-retry-all-errors" > /dev/null [ $? == 0 ] && extra_params="--retry-all-errors ${extra_params}" status=$(/usr/bin/curl -I -o /dev/null -skL $extra_params --connect-timeout ${timeout} --retry ${try} -w %{http_code} "$url") case "$status" in 204|\ 200) status=200 ;; esac echo $status } test_proxy() { result=0 status=$(test_url "https://www.google.com/generate_204" ${retry_num} ${connect_timeout}) if [ "$status" = "200" ]; then result=0 else status2=$(test_url "https://www.baidu.com" ${retry_num} ${connect_timeout}) if [ "$status2" = "200" ]; then result=1 else result=2 ping -c 3 -W 1 223.5.5.5 > /dev/null 2>&1 [ $? -eq 0 ] && { result=1 } fi fi echo $result } url_test_node() { result=0 local node_id=$1 local _type=$(echo $(config_n_get ${node_id} type) | tr 'A-Z' 'a-z') [ -n "${_type}" ] && { if [ "${_type}" == "socks" ]; then local _address=$(config_n_get ${node_id} address) local _port=$(config_n_get ${node_id} port) [ -n "${_address}" ] && [ -n "${_port}" ] && { local curlx="socks5h://${_address}:${_port}" local _username=$(config_n_get ${node_id} username) local _password=$(config_n_get ${node_id} password) [ -n "${_username}" ] && [ -n "${_password}" ] && curlx="socks5h://${_username}:${_password}@${_address}:${_port}" } else local _tmp_port=$(/usr/share/${CONFIG}/app.sh get_new_port 61080 tcp) /usr/share/${CONFIG}/app.sh run_socks flag="url_test_${node_id}" node=${node_id} bind=127.0.0.1 socks_port=${_tmp_port} config_file=url_test_${node_id}.json local curlx="socks5h://127.0.0.1:${_tmp_port}" fi sleep 1s # 兼容 curl 8.6 time_starttransfer 错误 local _cmd="-V 2>/dev/null | head -n 1 | awk '{print \$2}' | cut -d. -f1,2 | tr -d ' \\n'" local _curl="/usr/bin/curl" local curl_ver=$(lua_api "get_bin_version_cache(\"${_curl}\", \"${_cmd}\")") local curl_arg="-w %{http_code}:%{time_starttransfer} http://" [ "${curl_ver}" = "8.6" ] && curl_arg="-w %{http_code}:%{time_appconnect} https://" local chn_list=$(config_n_get @global[0] chn_list direct) local probeUrl="www.google.com/generate_204" [ "${chn_list}" = "proxy" ] && probeUrl="www.baidu.com" result=$(${_curl} --max-time 5 -o /dev/null -I -skL -x ${curlx} ${curl_arg}${probeUrl}) # 结束 SS 插件进程 local pid_file="/tmp/etc/${CONFIG}/url_test_${node_id}_plugin.pid" [ -s "$pid_file" ] && kill -9 "$(head -n 1 "$pid_file")" >/dev/null 2>&1 pgrep -af "url_test_${node_id}" | awk '! /test\.sh/{print $1}' | xargs kill -9 >/dev/null 2>&1 rm -rf /tmp/etc/${CONFIG}/*url_test_${node_id}*.* } echo $result } arg1=$1 shift case $arg1 in test_url) test_url $@ ;; url_test_node) url_test_node $@ ;; esac
294coder/Efficient-MIF
4,574
readmes/FeINFN.md
# Fourier-enhanced Implicit Neural Fusion Network for Multispectral and Hyperspectral Image Fusion <div align="center"> <p style="text-align: center"> <a href="https://scholar.google.com/citations?user=E5KO9XsAAAAJ&hl=en", style="font-size: 18px;">Yu-Jie Liang</a> &nbsp <a href="https://scholar.google.com/citations?user=pv61p_EAAAAJ&hl=en", style="font-size: 18px;">Zihan Cao</a> &nbsp <a href="https://scholar.google.com/citations?user=JZag1WIAAAAJ&hl=en", style="font-size: 18px;"> Shangqi Deng </a> &nbsp <a style="font-size: 18px;"> Hong-Xia Dou </a> &nbsp <a href="https://liangjiandeng.github.io/", style="font-size: 18px;"> Liang-Jian Deng </a> <br> <a style="font-size: 16px;"> University of Electronic Science & Technology of China </a> </p> </div> [![arXiv](https://img.shields.io/badge/arXiv-2404.09293-b31b1b.svg)](https://arxiv.org/abs/2404.15174) abtract: Recently, implicit neural representations (INR) have made significant strides in various vision-related domains, providing a novel solution for Multispectral and Hyperspectral Image Fusion (MHIF) tasks. However, INR is prone to losing high-frequency information and is confined to the lack of global perceptual capabilities. To address these issues, this paper introduces a Fourier-enhanced Implicit Neural Fusion Network (FeINFN) specifically designed for MHIF task, targeting the following phenomena: \textit{The Fourier amplitudes of the HR-HSI latent code and LR-HSI are remarkably similar; however, their phases exhibit different patterns.} In FeINFN, we innovatively propose a spatial and frequency implicit fusion function (Spa-Fre IFF), helping INR capture high-frequency information and expanding the receptive field. Besides, a new decoder employing a complex Gabor wavelet activation function, called Spatial-Frequency Interactive Decoder (SFID), is invented to enhance the interaction of INR features. Especially, we further theoretically prove that the Gabor wavelet activation possesses a time-frequency tightness property that favors learning the optimal bandwidths in the decoder. Experiments on two benchmark MHIF datasets verify the state-of-the-art (SOTA) performance of the proposed method, both visually and quantitatively. Also, ablation studies demonstrate the mentioned contributions. <html> <body> <div class="image-container" style="text-align: center;"> <img src="../teasers/FeINFN.png" alt="Image 1" width="100%"> </div> </body> </html> # Model We implement FeINFN with Pytorch and you can find it at [`model/FeINFN.py`](../model/FeINFN.py). ## Traning To train the model, running the following commands: ```shell CUDA_VISIBLE_DEVICES="0" \ NCCL_P2P_LEVEL="NVL" \ NCCL_P2P_DISABLE="1" \ NCCL_IB_DISABLE="1" \ OMP_NUM_THREADS="6" \ accelerate launch \ --config_file configs/huggingface/accelerate.yaml \ accelerate_main.py \ --proj_name FeINFN \ --arch FeINFN \ --dataset <dataset_name> \ --num_worker 0 -e 2000 -b 4 --aug_probs 0. 0. --loss l1ssim --grad_accum_steps 2 \ --checkpoint_every_n 20 --val_n_epoch 20 \ --comment "FeINFN config on <dataset_name> dataset model" \ --log_metric \ --logger_on \ ``` > check the `model/__init__.py` if the FeINFN network is not registered. Checkpoints, running, and Tensorboard logs will be saved at `log_file/`. ## Testing You can refer to the testing script [`torch_inference_on_sharpening.py`](../torch_inference_on_sharpening.py) to test the model. To test the metrics, please see the main guidance in [`README.md`](../README.md). CAVE pretrained model is released [here](https://pan.baidu.com/s/1JMRidVmoXZ6tKbtsGmG__w?pwd=mtef) (BaiduYun Disk). For sharpening tasks (including pansharpening and HMIF tasks), you simply test the metrics in Matlab: ``` matlab cd Pansharpening_Hyper_SR_Matlab_Test_Package %% when testing the reduced-resolution metrics on MHIF tasks % Args: % path: the saved fused image `.mat` file, find it in `visualized_img/` % ratio: upscale ratio, e.g., 4 % full_res: we keep it to 0, not changed % const: max value of the dataset (CAVE x4: 1, Harvard x4: 1, CAVE x8: 1, Harvard x8: 1) analysis_ref_batched_images(path, ratio, full_res, const) ``` # Citation If you find this work useful, please consider citing: ```bibtex @article{liang2024fourier, title={Fourier-enhanced Implicit Neural Fusion Network for Multispectral and Hyperspectral Image Fusion}, author={Liang, Yu-Jie and Cao, Zihan and Deng, ShangQi and Dou Hong-Xia and Deng, Liang-Jian}, journal={arXiv preprint arXiv:2404.15174}, year={2024} } ```
294coder/Efficient-MIF
3,890
readmes/lformer.md
# Linearly-evolved Transformer for Pan-sharpening <div align="center"> <p style="text-align: center"> <a style="font-size: 18px;"> JunMing Hou* </a> &nbsp <a href="https://scholar.google.com/citations?user=pv61p_EAAAAJ&hl=en", style="font-size: 18px;">Zihan Cao*</a> &nbsp <a style="font-size: 18px;"> Naishan Zheng </a> &nbsp <a style="font-size: 18px;"> Xuan Li </a> &nbsp <a style="font-size: 18px;"> Xiaoyu Chen </a> &nbsp <br> <a style="font-size: 18px;"> XinYang Li </a> &nbsp <a style="font-size: 18px;"> Xiaofeng Cong</a> &nbsp <a style="font-size: 18px;"> Man Zhou </a> &nbsp <a style="font-size: 18px;"> Danfeng Hong </a> &nbsp <br> <a style="font-size: 16px;"> University of Electronic Science Technology of China </a> <br> <a style="font-size: 16px;"> Southeast University </a> <br> <a style="font-size: 16px;"> University of Science and Technology </a> </p> </div> [![arXiv](https://img.shields.io/badge/arXiv-2404.12804-b31b1b.svg)](https://arxiv.org/abs/2404.12804) # Fast testing We provide [pretrained weights](https://pan.baidu.com/s/1keK5eAIrZcPPgoEr8bcW5A?pwd=y2t9) and a fast testing script to test the performance of our model. To run the testing script, please refer to `torch_inference_on_sharpening.py` and adapt following steps: 1. modify the `path` for datset; 2. change the `dataset_type`; 3. set `full_res` to the full resolution or reduced resolution of datasets (for pansharpening). 4. change the yaml file for configurate the model. For LFormer, it's in `configs/lformer_config.yaml`. ```yaml network_configs: lformer: pan_dim: 1 lms_dim: 4 # 4 for GF2, 8 for WV3, 31 for CAVE x4 attn_dim: 64 hp_dim: 64 n_stage: 5 patch_merge: yes crop_batch_size: 64 patch_size_list: [16, 64, 64] scale: 4 ``` To test the metrics, please see the main guidance in [`README.md`](../README.md). For sharpening tasks (including pansharpening and HMIF tasks), you simply test the metrics in Matlab: ``` matlab cd Pansharpening_Hyper_SR_Matlab_Test_Package %% when testing the reduced-resolution metrics on pansharpening or HMIF tasks % Args: % path: the saved fused image `.mat` file, find it in `visualized_img/` % ratio: upscale ratio, e.g., 4 % full_res: we keep it to 1, not changed % const: max value of the dataset (WV3: 2047, GF2: 1023, CAVE x8: 1, Harvard x8: 1) analysis_ref_batched_images(path, ratio, full_res, const) %% when testing full-resolution metrics on pansharpening datasets % Args: % path: the saved fused image `.mat` file, find it in `visualized_img/` % ratio: upscale ratio, e.g., 4 % sensor: the sensor name ('WV3', 'GF2', 'CAVE', 'Harvard') analysis_unref_batched_images(path, ratio, sensor) ``` # Train You can train the LFormer model by run the commands: ```shell CUDA_VISIBLE_DEVICES="0" \ NCCL_P2P_LEVEL="NVL" \ NCCL_P2P_DISABLE="1" \ NCCL_IB_DISABLE="1" \ OMP_NUM_THREADS="6" \ accelerate launch \ --config_file configs/huggingface/accelerate.yaml \ accelerate_main.py \ --proj_name LFormer \ --arch LFormer \ --dataset <dataset_name> \ --num_worker 6 -e 800 -b 32 --aug_probs 0. 0. --loss l1ssim --grad_accum_steps 2 \ --checkpoint_every_n 20 --val_n_epoch 20 \ --comment "LFormer config on wv3 dataset model" \ --log_metric \ --logger_on \ ``` > Other DDP training (or using deepspeed, please see Huggingface Accelerate documenation). # Citation If you find this code useful for your research, please consider citing: ``` @inproceedings{ hou2024linearlyevolved, title={Linearly-evolved Transformer for Pan-sharpening}, author={Junming Hou and Zihan Cao and Naishan Zheng and Xuan Li and Xiaoyu Chen and Xinyang Liu and Xiaofeng Cong and Danfeng Hong and Man Zhou}, booktitle={ACM Multimedia 2024}, year={2024}, url={https://openreview.net/forum?id=pCxZTmGr4O} } ```
294coder/Efficient-MIF
5,435
readmes/le-mamba.md
# A Novel State Space Model with Local Enhancement and State Sharing for Image Fusion <div align="center"> <p style="text-align: center"> <a href="https://scholar.google.com/citations?user=pv61p_EAAAAJ&hl=en", style="font-size: 18px;">Zihan Cao</a> &nbsp <a href="https://xiaoxiao-woo.github.io/", style="font-size: 18px;"> Xiao Wu </a> &nbsp <a href="https://liangjiandeng.github.io/", style="font-size: 18px;"> Liang-Jian Deng </a> <a style="font-size: 18px;"> Yu Zhong </a> <br> <a style="font-size: 16px;"> University of Electronic Science Technology of China </a> </p> </div> [![arXiv](https://img.shields.io/badge/arXiv-2404.09293-b31b1b.svg)](https://arxiv.org/abs/2404.09293) abtract: In image fusion tasks, images from different sources possess distinct characteristics. This has driven the development of numerous methods to explore better ways of fusing them while preserving their respective characteristics. Mamba, as a state space model, has emerged in the field of natural language processing. Recently, many studies have attempted to extend Mamba to vision tasks. However, due to the nature of images different from causal language sequences, the limited state capacity of Mamba weakens its ability to model image information. Additionally, the sequence modeling ability of Mamba is only capable of spatial information and cannot effectively capture the rich spectral information in images. Motivated by these challenges, we customize and improve the vision Mamba network designed for the image fusion task. Specifically, we propose the local-enhanced vision Mamba block, dubbed as LEVM. The LEVM block can improve local information perception of the network and simultaneously learn local and global spatial information. Furthermore, we propose the state sharing technique to enhance spatial details and integrate spatial and spectral information. Finally, the overall network is a multi-scale structure based on vision Mamba, called LE-Mamba. Extensive experiments show the proposed methods achieve state-of-the-art results on multispectral pansharpening and multispectral and hyperspectral image fusion datasets, and demonstrate the effectiveness of the proposed approach. <html> <body> <div class="image-container" style="text-align: center;"> <img src="figs/le-mamba-teaser.png" alt="Image 1" width="100%"> </div> </body> </html> # Model We implement LE-Mamba with Pytorch and you can find it at [`model/LEMamba.py`](../model/LEMamba.py). ## Traning To train the model, you should first compile the `select_scan` operator by running the following command: ```shell cd model/module/mamba_scan pip install -e . ``` Then, you should run the following command to test the model if it is correctly installed: ```shell pytest model/test/LEMamba_test.py ``` If the test is passed, you can start training the model by running the following commands: ```shell CUDA_VISIBLE_DEVICES="0" \ NCCL_P2P_LEVEL="NVL" \ NCCL_P2P_DISABLE="1" \ NCCL_IB_DISABLE="1" \ OMP_NUM_THREADS="6" \ accelerate launch \ --config_file configs/huggingface/accelerate.yaml \ accelerate_main.py \ --proj_name LE-Mamba \ --arch LEMamba \ --dataset <dataset_name> \ --num_worker 6 -e 800 -b 32 --aug_probs 0. 0. --loss l1ssim --grad_accum_steps 2 \ --checkpoint_every_n 20 --val_n_epoch 20 \ --comment "LE-Mamba config on wv3 dataset model" \ --log_metric \ --logger_on \ ``` > check the `model/__init__.py` if the LE-Mamba network is not registered. Checkpoints, running, and Tensorboard logs will be saved at `log_file/`. ## Testing You can refer to the testing script [`torch_inference_on_sharpening.py`](../torch_inference_on_sharpening.py) to test the model. To test the metrics, please see the main guidance in [`README.md`](../README.md). For sharpening tasks (including pansharpening and HMIF tasks), you simply test the metrics in Matlab: ``` matlab cd Pansharpening_Hyper_SR_Matlab_Test_Package %% when testing the reduced-resolution metrics on pansharpening or HMIF tasks % Args: % path: the saved fused image `.mat` file, find it in `visualized_img/` % ratio: upscale ratio, e.g., 4 % full_res: we keep it to 1, not changed % const: max value of the dataset (WV3: 2047, GF2: 1023, CAVE x8: 1, Harvard x8: 1) analysis_ref_batched_images(path, ratio, full_res, const) %% when testing full-resolution metrics on pansharpening datasets % Args: % path: the saved fused image `.mat` file, find it in `visualized_img/` % ratio: upscale ratio, e.g., 4 % sensor: the sensor name ('WV3', 'GF2', 'CAVE', 'Harvard') analysis_unref_batched_images(path, ratio, sensor) ``` <!-- # Performances LE-Mamba reaches SOTA performances on widely-used Pansharpening and HMIF datasets. Here are some metrics: <html> <body> <div class="image-container" style="text-align: center;"> <img src="figs/le-mamba-pansharpening.png" alt="Image 1" width="100%"> </div> <div class="image-container" style="text-align: center;"> <img src="figs/le-mamba-HMIF.png" alt="Image 1" width="100%"> </div> </body> </html> --> # Citation If you find this work useful, please consider citing: ```bibtex @inproceedings{ cao2024novel, title={A novel state space model with local enhancement and state sharing for image fusion}, author={Cao, Zihan and Wu, Xiao and Deng, Liang-Jian and Zhong, Yu}, booktitle={ACM Multimedia 2024 (ACM MM)}, year={2024} } ```
2977094657/DidaAPI
10,585
services/dida_service.py
"""滴答清单API服务模块""" import uuid import time from typing import Optional, Dict, Any, List import httpx from utils import app_logger from core import config, db, urls from models import TasksResponse, TaskItem class DidaAPIService: """滴答清单API服务类""" def __init__(self): self.request_config = config.get('request_config', {}) self.client = httpx.AsyncClient(timeout=self.request_config.get('timeout', 30.0)) # 从数据库获取当前活跃的认证会话 self.current_session = None self._load_active_session() def _load_active_session(self): """从数据库加载活跃的认证会话""" try: # 从数据库获取最新的活跃会话 session_data = db.get_latest_active_session() if session_data: self.current_session = { 'session_id': session_data['session_id'], 'auth_token': session_data['token'], 'csrf_token': session_data['csrf_token'], 'is_active': session_data['is_active'] } app_logger.info(f"已从数据库恢复认证会话: {session_data['session_id']}") else: app_logger.info("数据库中没有找到活跃的认证会话") except Exception as e: app_logger.error(f"加载认证会话失败: {e}") def set_auth_session(self, auth_token: str, csrf_token: str) -> str: """设置认证会话""" session_id = str(uuid.uuid4()) self.current_session = { 'session_id': session_id, 'auth_token': auth_token, 'csrf_token': csrf_token, 'is_active': True } # 保存到数据库 db.save_user_session({ 'session_id': session_id, 'token': auth_token, 'csrf_token': csrf_token, 'is_active': True }) app_logger.info(f"设置认证会话成功: {session_id}") return session_id def get_session_status(self) -> Dict[str, Any]: """获取当前会话状态""" if self.current_session: return { "has_session": True, "session_id": self.current_session.get('session_id'), "is_active": self.current_session.get('is_active', False) } else: return { "has_session": False, "session_id": None, "is_active": False } def _get_auth_headers(self) -> Dict[str, str]: """获取认证请求头""" if not self.current_session: raise ValueError("未设置认证会话,请先登录") # 生成traceid traceid = f"{int(time.time() * 1000):x}{uuid.uuid4().hex[:8]}" headers = { 'Accept': 'application/json, text/plain, */*', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7', 'Cache-Control': 'no-cache', 'Origin': 'https://dida365.com', 'Pragma': 'no-cache', 'Referer': 'https://dida365.com/', 'Sec-Ch-Ua': '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"', 'Sec-Ch-Ua-Mobile': '?0', 'Sec-Ch-Ua-Platform': '"Windows"', 'Sec-Fetch-Dest': 'empty', 'Sec-Fetch-Mode': 'cors', 'Sec-Fetch-Site': 'same-site', 'User-Agent': self.request_config.get('user_agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'), 'X-Csrftoken': self.current_session['csrf_token'], 'X-Device': self.request_config.get('device_info', '{}'), 'Hl': self.request_config.get('language', 'zh_CN'), 'X-Tz': self.request_config.get('timezone', 'Asia/Shanghai'), 'Traceid': traceid } return headers def _get_auth_cookies(self) -> Dict[str, str]: """获取认证cookies""" if not self.current_session: raise ValueError("未设置认证会话,请先登录") cookies = { 't': self.current_session['auth_token'], '_csrf_token': self.current_session['csrf_token'] } return cookies async def get_all_tasks(self) -> dict: """ 获取所有任务 Returns: dict: 原始响应数据 """ try: if not self.current_session: return {"error": "no_auth_session", "message": "未设置认证会话,请先登录"} # 使用统一的URL构建函数 url = urls.build_dida_api_url(urls.DIDA_TASK_APIS["get_all_tasks"]) # 获取认证头和cookies headers = self._get_auth_headers() cookies = self._get_auth_cookies() app_logger.info(f"请求获取所有任务: {url}") app_logger.debug(f"请求头: {headers}") # 发送请求 response = await self.client.get(url, headers=headers, cookies=cookies) # 记录响应信息 app_logger.info(f"任务响应状态码: {response.status_code}") app_logger.debug(f"任务响应头: {dict(response.headers)}") if response.status_code == 200: # 解析响应数据 response_data = response.json() app_logger.info(f"成功获取任务数据,响应长度: {len(str(response_data))}") app_logger.debug(f"任务响应数据: {response_data}") # 直接返回原始响应 return response_data else: app_logger.error(f"获取任务失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"获取任务时发生错误: {e}") return {"error": str(e)} async def get_completed_tasks(self, to: Optional[str] = None, status: str = "Completed") -> dict: """ 获取已完成或已放弃的任务(支持分页) Args: to: 分页参数,使用上次响应最后一个任务的completedTime字段 如果为None,则获取第一页 后续请求使用上次响应最后一个任务的completedTime字段(原始格式) 原始格式:2025-03-15T13:30:54.000+0000 API格式:2025-03-15 13:30:54 status: 任务状态,支持以下值: - "Completed": 已完成的任务 - "Abandoned": 已放弃的任务 Returns: dict: 原始响应数据,包含任务列表 Note: 分页机制: - 第一次请求:不传to参数 - 后续请求:使用上次响应最后一个任务的completedTime字段作为to参数 - completedTime原始格式:2025-03-15T13:30:54.000+0000 - API需要格式:2025-03-15 13:30:54 - URL示例:https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Completed """ try: if not self.current_session: return {"error": "no_auth_session", "message": "未设置认证会话,请先登录"} # 构建URL base_url = urls.build_dida_api_url(urls.DIDA_TASK_APIS["get_completed_tasks"]) # 构建查询参数 params = { "from": "", # 固定为空 "status": status # 支持Completed或Abandoned } # 如果提供了to参数,则添加到查询参数中 if to: # 将completedTime格式转换为滴答清单API需要的格式 # 从 2025-03-15T13:30:54.000+0000 转换为 2025-03-15 13:30:54 formatted_to = to.replace('T', ' ').replace('.000+0000', '') params["to"] = formatted_to # 第一次请求不添加to参数 # 获取认证头和cookies headers = self._get_auth_headers() cookies = self._get_auth_cookies() app_logger.info(f"请求获取已完成任务: {base_url}") app_logger.info(f"查询参数: {params}") app_logger.debug(f"请求头: {headers}") # 发送请求 response = await self.client.get(base_url, headers=headers, cookies=cookies, params=params) # 记录响应信息 app_logger.info(f"已完成任务响应状态码: {response.status_code}") app_logger.debug(f"已完成任务响应头: {dict(response.headers)}") if response.status_code == 200: # 解析响应数据 response_data = response.json() task_count = len(response_data) if isinstance(response_data, list) else 0 app_logger.info(f"成功获取已完成任务数据,任务数量: {task_count}") app_logger.debug(f"已完成任务响应数据: {response_data}") # 直接返回原始响应 return response_data else: app_logger.error(f"获取已完成任务失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"获取已完成任务时发生错误: {e}") return {"error": str(e)} async def get_trash_tasks(self, limit: int = 50, task_type: int = 1) -> dict: """ 获取垃圾桶中的任务 Args: limit: 每页任务数量,默认50 task_type: 任务类型,默认1 Returns: dict: 原始响应数据,包含垃圾桶任务列表 Note: 响应格式: { "tasks": [...], # 任务列表 "next": 0 # 下一页标识 } """ try: if not self.current_session: return {"error": "no_auth_session", "message": "未设置认证会话,请先登录"} # 构建URL base_url = urls.build_dida_api_url(urls.DIDA_TASK_APIS["get_trash_tasks"]) # 构建查询参数 params = { "limit": limit, "type": task_type } # 获取认证头和cookies headers = self._get_auth_headers() cookies = self._get_auth_cookies() app_logger.info(f"请求获取垃圾桶任务: {base_url}") app_logger.info(f"查询参数: {params}") app_logger.debug(f"请求头: {headers}") # 发送请求 response = await self.client.get(base_url, headers=headers, cookies=cookies, params=params) # 记录响应信息 app_logger.info(f"垃圾桶任务响应状态码: {response.status_code}") app_logger.debug(f"垃圾桶任务响应头: {dict(response.headers)}") if response.status_code == 200: # 解析响应数据 response_data = response.json() task_count = len(response_data.get('tasks', [])) if isinstance(response_data, dict) else 0 app_logger.info(f"成功获取垃圾桶任务数据,任务数量: {task_count}") app_logger.debug(f"垃圾桶任务响应数据: {response_data}") # 直接返回原始响应 return response_data else: app_logger.error(f"获取垃圾桶任务失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"获取垃圾桶任务时发生错误: {e}") return {"error": str(e)} async def close(self): """关闭HTTP客户端""" await self.client.aclose() # 全局滴答清单API服务实例 dida_service = DidaAPIService()
2977094657/DidaAPI
3,363
services/statistics_service.py
"""统计服务模块""" import httpx from utils import app_logger from core import urls class StatisticsService: """统计服务类""" def __init__(self): self.client = httpx.AsyncClient(timeout=30.0) def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict: """构建认证请求头""" return { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36', 'Accept': 'application/json, text/plain, */*', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8', 'Content-Type': 'application/json', 'X-Requested-With': 'XMLHttpRequest', 'X-Tz': 'Asia/Shanghai', } def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict: """构建认证cookies""" return { 't': auth_token, '_csrf_token': csrf_token } async def get_user_ranking(self, auth_token: str, csrf_token: str) -> dict: """获取用户排名统计,直接返回原始响应""" try: url = urls.build_dida_api_url(urls.DIDA_STATISTICS_APIS["user_ranking"]).replace('/v2/', '/v3/') headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def get_general_statistics(self, auth_token: str, csrf_token: str) -> dict: """获取通用统计信息,直接返回原始响应""" try: url = urls.build_dida_api_url(urls.DIDA_STATISTICS_APIS["general_statistics"]) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def get_task_statistics(self, auth_token: str, csrf_token: str, start_date: str, end_date: str) -> dict: """获取任务统计信息,直接返回原始响应""" try: endpoint = f"{urls.DIDA_STATISTICS_APIS['task_statistics']}/{start_date}/{end_date}" url = urls.build_dida_api_url(endpoint) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: return response.json() else: return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: return {"error": str(e)} async def close(self): """关闭HTTP客户端""" await self.client.aclose() # 全局统计服务实例 statistics_service = StatisticsService()
294coder/Efficient-MIF
1,073
Pansharpening_Hyper_SR_Matlab_Test_Package/quality_assess.m
function [mean_psnr, mean_ssim] = quality_assess(imagery1, imagery2) %========================================================================== % Evaluates the quality assessment indices for two tensors. % % Syntax: % [mpsnr, mssim] = quality_access(imagery1, imagery2) % % Input: % imagery1 - the reference tensor % imagery2 - the target tensor % NOTE: the tensor is a I1*I2*...*IN array and DYNAMIC RANGE [0, 255]. % Output: % mpsnr - Peak Signal-to-Noise Ratio % mssim - Structure SIMilarity %========================================================================== Nway = size(imagery1); if length(Nway)>3 imagery1 = reshape(imagery1,Nway(1),Nway(2),[]); imagery2 = reshape(imagery2,Nway(1),Nway(2),[]); end psnr = zeros(prod(Nway(3:end)),1); ssim = psnr; for ii = 1:prod(Nway(3:end)) psnr(ii) = psnr_index(imagery1(:, :, ii), imagery2(:, :, ii)); % 因为ssim_index是邓尚琦给的代码,需要乘255 ssim(ii) = ssim_index(imagery1(:, :, ii)*255, imagery2(:, :, ii)*255); end mean_psnr = mean(psnr); mean_ssim = mean(ssim); %out = [mean(psnr), mean(ssim)];
2977094657/DidaAPI
2,429
services/user_service.py
"""用户信息服务模块""" import httpx from core import urls from utils import app_logger class UserService: """用户信息服务类""" def __init__(self): self.client = httpx.AsyncClient(timeout=30.0) def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict: """构建认证headers""" return { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36', 'Accept': 'application/json, text/plain, */*', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8', 'Content-Type': 'application/json;charset=UTF-8', 'X-Requested-With': 'XMLHttpRequest', 'x-csrftoken': csrf_token, 'Referer': 'https://dida365.com/', 'Origin': 'https://dida365.com' } def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict: """构建认证cookies""" return { 't': auth_token, '_csrf_token': csrf_token } async def get_user_profile(self, auth_token: str, csrf_token: str) -> dict: """ 获取用户信息,直接返回原始响应 Args: auth_token: 认证令牌 csrf_token: CSRF令牌 Returns: dict: 原始API响应 """ try: url = urls.build_dida_api_url(urls.DIDA_AUTH_APIS["user_profile"]) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) app_logger.info(f"请求获取用户信息: {url}") response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: response_data = response.json() app_logger.info(f"成功获取用户信息,用户名: {response_data.get('username', 'N/A')}") # 直接返回原始响应 return response_data else: app_logger.error(f"获取用户信息失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"获取用户信息时发生错误: {e}") return {"error": str(e)} async def close(self): """关闭HTTP客户端""" await self.client.aclose() # 全局用户服务实例 user_service = UserService()
2977094657/DidaAPI
23,614
services/export_service.py
"""任务导出服务""" import io from typing import Dict, List, Any, Optional from datetime import datetime import pandas as pd from utils import app_logger from services.dida_service import dida_service from services.pomodoro_service import pomodoro_service from core import urls class ExportService: """任务导出服务类""" def __init__(self): self.dida_service = dida_service async def export_tasks_to_excel(self) -> Dict[str, Any]: """ 导出所有任务到Excel文件 Returns: dict: 包含Excel文件内容和元数据的响应 """ try: app_logger.info("开始导出任务到Excel") # 获取所有任务数据 all_tasks_data = await self._get_all_tasks_data() completed_tasks_data = await self._get_completed_tasks_data() abandoned_tasks_data = await self._get_abandoned_tasks_data() trash_tasks_data = await self._get_trash_tasks_data() if not all_tasks_data and not completed_tasks_data and not abandoned_tasks_data and not trash_tasks_data: return {"error": "无法获取任务数据"} # 创建Excel文件 excel_buffer = io.BytesIO() with pd.ExcelWriter(excel_buffer, engine='openpyxl') as writer: # 处理全部任务 if all_tasks_data: all_tasks_df = self._process_all_tasks(all_tasks_data) if not all_tasks_df.empty: all_tasks_df.to_excel(writer, sheet_name='全部任务', index=False) app_logger.info(f"全部任务工作表创建完成,共 {len(all_tasks_df)} 条记录") # 处理已完成任务 if completed_tasks_data: completed_tasks_df = self._process_completed_tasks(completed_tasks_data) if not completed_tasks_df.empty: completed_tasks_df.to_excel(writer, sheet_name='已完成任务', index=False) app_logger.info(f"已完成任务工作表创建完成,共 {len(completed_tasks_df)} 条记录") # 处理放弃任务 if abandoned_tasks_data: abandoned_tasks_df = self._process_abandoned_tasks(abandoned_tasks_data) if not abandoned_tasks_df.empty: abandoned_tasks_df.to_excel(writer, sheet_name='放弃任务', index=False) app_logger.info(f"放弃任务工作表创建完成,共 {len(abandoned_tasks_df)} 条记录") # 处理垃圾桶任务 if trash_tasks_data: trash_tasks_df = self._process_trash_tasks(trash_tasks_data) if not trash_tasks_df.empty: trash_tasks_df.to_excel(writer, sheet_name='垃圾桶任务', index=False) app_logger.info(f"垃圾桶任务工作表创建完成,共 {len(trash_tasks_df)} 条记录") excel_buffer.seek(0) # 生成文件名 timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") filename = f"滴答清单任务导出_{timestamp}.xlsx" app_logger.info(f"Excel文件生成完成: {filename}") return { "filename": filename, "content_type": "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", "content": excel_buffer.getvalue(), "size": len(excel_buffer.getvalue()) } except Exception as e: app_logger.error(f"导出任务到Excel时发生错误: {e}") return {"error": str(e)} async def export_focus_records_to_excel(self) -> Dict[str, Any]: """ 导出专注记录到Excel文件 Returns: dict: 包含Excel文件内容和元数据的响应 """ try: app_logger.info("开始导出专注记录到Excel") # 获取专注记录数据 focus_timeline_data = await self._get_all_focus_timeline_data() if not focus_timeline_data: return {"error": "无法获取专注记录数据"} # 创建Excel文件 excel_buffer = io.BytesIO() with pd.ExcelWriter(excel_buffer, engine='openpyxl') as writer: # 处理专注记录时间线 if focus_timeline_data: focus_timeline_df = self._process_focus_timeline(focus_timeline_data) if not focus_timeline_df.empty: focus_timeline_df.to_excel(writer, sheet_name='专注记录时间线', index=False) app_logger.info(f"专注记录时间线工作表创建完成,共 {len(focus_timeline_df)} 条记录") excel_buffer.seek(0) # 生成文件名 timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") filename = f"滴答清单专注记录导出_{timestamp}.xlsx" app_logger.info(f"专注记录Excel文件生成完成: {filename}") return { "filename": filename, "content_type": "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet", "content": excel_buffer.getvalue(), "size": len(excel_buffer.getvalue()) } except Exception as e: app_logger.error(f"导出专注记录到Excel时发生错误: {e}") return {"error": str(e)} async def _get_all_tasks_data(self) -> Optional[Dict]: """获取所有任务数据""" try: result = await self.dida_service.get_all_tasks() if result and 'error' not in result: return result return None except Exception as e: app_logger.error(f"获取所有任务数据失败: {e}") return None async def _get_completed_tasks_data(self) -> Optional[List]: """获取已完成任务数据(分页获取所有数据)""" try: all_completed_tasks = [] to = None page_count = 0 while True: app_logger.info(f"获取已完成任务第 {page_count + 1} 页,to参数: {to}") result = await self.dida_service.get_completed_tasks(to, "Completed") if not result or 'error' in result: app_logger.warning(f"获取已完成任务第 {page_count + 1} 页失败: {result}") break if isinstance(result, list) and len(result) > 0: all_completed_tasks.extend(result) app_logger.info(f"第 {page_count + 1} 页获取到 {len(result)} 条已完成任务") # 获取最后一个任务的completedTime作为下次分页参数 last_task = result[-1] to = last_task.get('completedTime') if not to: app_logger.info("最后一个任务没有completedTime,停止分页") break page_count += 1 # 如果返回的任务数少于50条,说明已经是最后一页 if len(result) < 50: app_logger.info("已获取到最后一页已完成任务") break else: app_logger.info("没有更多已完成任务数据") break app_logger.info(f"已完成任务分页获取完成,共获取 {len(all_completed_tasks)} 条记录,分 {page_count} 页") return all_completed_tasks if all_completed_tasks else None except Exception as e: app_logger.error(f"获取已完成任务数据失败: {e}") return None async def _get_abandoned_tasks_data(self) -> Optional[List]: """获取放弃任务数据(分页获取所有数据)""" try: all_abandoned_tasks = [] to = None page_count = 0 while True: app_logger.info(f"获取放弃任务第 {page_count + 1} 页,to参数: {to}") result = await self.dida_service.get_completed_tasks(to, "Abandoned") if not result or 'error' in result: app_logger.warning(f"获取放弃任务第 {page_count + 1} 页失败: {result}") break if isinstance(result, list) and len(result) > 0: all_abandoned_tasks.extend(result) app_logger.info(f"第 {page_count + 1} 页获取到 {len(result)} 条放弃任务") # 获取最后一个任务的completedTime作为下次分页参数 last_task = result[-1] to = last_task.get('completedTime') if not to: app_logger.info("最后一个任务没有completedTime,停止分页") break page_count += 1 # 如果返回的任务数少于50条,说明已经是最后一页 if len(result) < 50: app_logger.info("已获取到最后一页放弃任务") break else: app_logger.info("没有更多放弃任务数据") break app_logger.info(f"放弃任务分页获取完成,共获取 {len(all_abandoned_tasks)} 条记录,分 {page_count} 页") return all_abandoned_tasks if all_abandoned_tasks else None except Exception as e: app_logger.error(f"获取放弃任务数据失败: {e}") return None async def _get_trash_tasks_data(self) -> Optional[Dict]: """获取垃圾桶任务数据""" try: result = await self.dida_service.get_trash_tasks() if result and 'error' not in result: return result return None except Exception as e: app_logger.error(f"获取垃圾桶任务数据失败: {e}") return None async def _get_all_focus_timeline_data(self) -> Optional[List]: """获取所有专注记录时间线数据(分页获取所有数据)""" try: all_focus_records = [] to_timestamp = None page_count = 0 max_pages = 100 # 防止无限循环 # 获取认证信息 current_session = self.dida_service.current_session if not current_session: app_logger.error("未找到认证会话") return None auth_token = current_session['auth_token'] csrf_token = current_session['csrf_token'] while page_count < max_pages: app_logger.info(f"获取专注记录第 {page_count + 1} 页,to_timestamp: {to_timestamp}") result = await pomodoro_service.get_focus_timeline(auth_token, csrf_token, to_timestamp) if not result or 'error' in result: app_logger.warning(f"获取专注记录第 {page_count + 1} 页失败: {result}") break # 检查是否有数据 if not isinstance(result, list) or len(result) == 0: app_logger.info("没有更多专注记录数据") break all_focus_records.extend(result) app_logger.info(f"第 {page_count + 1} 页获取到 {len(result)} 条专注记录") # 获取最后一条记录的startTime作为下次分页参数 if len(result) > 0: last_record = result[-1] start_time = last_record.get('startTime') if start_time: # 转换时间格式用于下次请求 to_timestamp = pomodoro_service._convert_time_to_timestamp(start_time) else: app_logger.info("最后一条记录没有startTime,停止分页") break else: break page_count += 1 # 如果返回的记录数少于31条(通常每页31条),说明已经是最后一页 if len(result) < 31: app_logger.info("已获取到最后一页专注记录") break app_logger.info(f"专注记录分页获取完成,共获取 {len(all_focus_records)} 条记录,分 {page_count} 页") return all_focus_records if all_focus_records else None except Exception as e: app_logger.error(f"获取专注记录时间线数据失败: {e}") return None def _process_all_tasks(self, data: Dict) -> pd.DataFrame: """处理全部任务数据""" try: tasks = data.get('syncTaskBean', {}).get('update', []) projects = {p['id']: p['name'] for p in data.get('projectProfiles', [])} processed_tasks = [] for task in tasks: processed_task = self._flatten_task(task, projects) processed_tasks.append(processed_task) return pd.DataFrame(processed_tasks) except Exception as e: app_logger.error(f"处理全部任务数据失败: {e}") return pd.DataFrame() def _process_completed_tasks(self, data: List) -> pd.DataFrame: """处理已完成任务数据""" try: processed_tasks = [] for task in data: processed_task = self._flatten_task(task, {}) processed_tasks.append(processed_task) return pd.DataFrame(processed_tasks) except Exception as e: app_logger.error(f"处理已完成任务数据失败: {e}") return pd.DataFrame() def _process_abandoned_tasks(self, data: List) -> pd.DataFrame: """处理放弃任务数据""" try: processed_tasks = [] for task in data: processed_task = self._flatten_task(task, {}) processed_tasks.append(processed_task) return pd.DataFrame(processed_tasks) except Exception as e: app_logger.error(f"处理放弃任务数据失败: {e}") return pd.DataFrame() def _process_trash_tasks(self, data: Dict) -> pd.DataFrame: """处理垃圾桶任务数据""" try: tasks = data.get('tasks', []) processed_tasks = [] for task in tasks: processed_task = self._flatten_task(task, {}) processed_tasks.append(processed_task) return pd.DataFrame(processed_tasks) except Exception as e: app_logger.error(f"处理垃圾桶任务数据失败: {e}") return pd.DataFrame() def _process_focus_timeline(self, data: List) -> pd.DataFrame: """处理专注记录时间线数据 - 紧凑型展示""" try: processed_records = [] for record in data: # 为每个专注会话创建一条紧凑记录 compact_record = self._create_compact_focus_record(record) processed_records.append(compact_record) return pd.DataFrame(processed_records) except Exception as e: app_logger.error(f"处理专注记录时间线数据失败: {e}") return pd.DataFrame() def _flatten_task(self, task: Dict, projects: Dict) -> Dict: """展平任务数据,包含所有字段""" try: flattened = { # 基本信息 '任务ID': task.get('id', ''), '任务标题': task.get('title', ''), '任务内容': task.get('content', ''), '任务描述': task.get('desc', ''), '项目ID': task.get('projectId', ''), '项目名称': projects.get(task.get('projectId', ''), ''), '排序顺序': task.get('sortOrder', 0), # 状态和优先级 '任务状态': self._get_status_text(task.get('status', 0)), '状态代码': task.get('status', 0), '优先级': task.get('priority', 0), '完成进度': task.get('progress', 0), '删除状态': task.get('deleted', 0), # 时间相关 '创建时间': task.get('createdTime', ''), '修改时间': task.get('modifiedTime', ''), '开始日期': task.get('startDate', ''), '截止日期': task.get('dueDate', ''), '置顶时间': task.get('pinnedTime', ''), '完成时间': task.get('completedTime', ''), '删除时间': task.get('deletedTime', ''), # 时区和时间设置 '时区': task.get('timeZone', ''), '是否浮动时间': task.get('isFloating', False), '是否全天任务': task.get('isAllDay', False), # 重复设置 '重复任务ID': task.get('repeatTaskId', ''), '重复标志': task.get('repeatFlag', ''), '重复来源': task.get('repeatFrom', ''), '首次重复日期': task.get('repeatFirstDate', ''), # 提醒设置 '提醒设置': task.get('reminder', ''), '提醒列表': str(task.get('reminders', [])), '排除日期': str(task.get('exDate', [])), # 层级关系 '父任务ID': task.get('parentId', ''), '子任务ID列表': str(task.get('childIds', [])), # 其他属性 '标签列表': str(task.get('tags', [])), '子项目': str(task.get('items', [])), '附件数量': len(task.get('attachments', [])), '评论数量': task.get('commentCount', 0), '列ID': task.get('columnId', ''), '类型': task.get('kind', ''), '图片模式': task.get('imgMode', 0), # 创建者和删除者 '创建者ID': task.get('creator', 0), '删除者ID': task.get('deletedBy', 0), # 版本控制 '实体标签': task.get('etag', ''), # 专注相关 '番茄钟摘要': str(task.get('pomodoroSummaries', [])), '专注摘要': str(task.get('focusSummaries', [])), # 附件详情 '附件详情': str(task.get('attachments', [])), } return flattened except Exception as e: app_logger.error(f"展平任务数据失败: {e}") return {} def _get_status_text(self, status_code: int) -> str: """获取状态文本描述""" status_map = { 0: '未完成', 1: '进行中', 2: '已完成', -1: '已删除' } return status_map.get(status_code, f'未知状态({status_code})') def _create_compact_focus_record(self, record: Dict) -> Dict: """创建紧凑型专注记录""" try: from datetime import datetime, timedelta # 基本信息 session_id = record.get('id', '') session_start = record.get('startTime', '') session_end = record.get('endTime', '') pause_duration = record.get('pauseDuration', 0) # 计算总时长 total_duration = 0 if session_start and session_end: try: start_time = datetime.fromisoformat(session_start.replace('Z', '+00:00')) end_time = datetime.fromisoformat(session_end.replace('Z', '+00:00')) total_duration = int((end_time - start_time).total_seconds()) except: total_duration = 0 # 获取任务信息 tasks = record.get('tasks', []) task_titles = [] project_names = [] for task in tasks: if task.get('title'): task_titles.append(task['title']) if task.get('projectName'): project_names.append(task['projectName']) main_task = '; '.join(set(task_titles)) # 去重 main_project = '; '.join(set(project_names)) # 去重 # 生成专注时间段描述 focus_timeline = self._generate_focus_timeline(tasks, pause_duration) # 生成暂停模式描述 pause_pattern = self._generate_pause_pattern(tasks, pause_duration) # 格式化会话时间 session_time_str = "" if session_start and session_end: try: start_dt = datetime.fromisoformat(session_start.replace('Z', '+00:00')) end_dt = datetime.fromisoformat(session_end.replace('Z', '+00:00')) session_time_str = f"{start_dt.strftime('%Y-%m-%d %H:%M')} - {end_dt.strftime('%H:%M')}" except: session_time_str = f"{session_start} - {session_end}" return { '会话ID': session_id, '会话时间': session_time_str, '总时长': self._format_duration(total_duration), '暂停时长': self._format_duration(pause_duration), '任务标题': main_task, '项目': main_project, '专注时间段': focus_timeline, '暂停模式': pause_pattern, '效率(%)': round((total_duration - pause_duration) / total_duration * 100, 1) if total_duration > 0 else 0, '时间段数量': len(tasks), '会话类型': record.get('type', ''), '实体标签': record.get('etag', '') } except Exception as e: app_logger.error(f"创建紧凑型专注记录失败: {e}") return {} def _generate_focus_timeline(self, tasks: List[Dict], total_pause_duration: int) -> str: """生成专注时间段描述""" try: from datetime import datetime if not tasks: return "无专注时间段" timeline_parts = [] for i, task in enumerate(tasks): start_time = task.get('startTime', '') end_time = task.get('endTime', '') if start_time and end_time: try: start_dt = datetime.fromisoformat(start_time.replace('Z', '+00:00')) end_dt = datetime.fromisoformat(end_time.replace('Z', '+00:00')) duration = int((end_dt - start_dt).total_seconds()) # 格式化时间段 time_part = f"{start_dt.strftime('%H:%M')}-{end_dt.strftime('%H:%M')}({self._format_duration(duration)})" timeline_parts.append(time_part) # 如果不是最后一个时间段,计算暂停时间 if i < len(tasks) - 1: next_task = tasks[i + 1] next_start = next_task.get('startTime', '') if next_start: try: next_start_dt = datetime.fromisoformat(next_start.replace('Z', '+00:00')) pause_duration = int((next_start_dt - end_dt).total_seconds()) if pause_duration > 0: timeline_parts.append(f"[暂停{self._format_duration(pause_duration)}]") except: timeline_parts.append("[暂停未知时长]") except: timeline_parts.append(f"时间段{i+1}(解析失败)") return " → ".join(timeline_parts) except Exception as e: app_logger.error(f"生成专注时间段描述失败: {e}") return "生成失败" def _generate_pause_pattern(self, tasks: List[Dict], total_pause_duration: int) -> str: """生成暂停模式描述""" try: if len(tasks) <= 1: return "无暂停" if total_pause_duration == 0 else f"总暂停{self._format_duration(total_pause_duration)}" pause_count = len(tasks) - 1 avg_pause = total_pause_duration // pause_count if pause_count > 0 else 0 if pause_count == 1: return f"暂停1次({self._format_duration(total_pause_duration)})" else: return f"暂停{pause_count}次(总计{self._format_duration(total_pause_duration)}, 平均{self._format_duration(avg_pause)})" except Exception as e: app_logger.error(f"生成暂停模式描述失败: {e}") return "分析失败" def _format_duration(self, seconds: int) -> str: """格式化时长显示""" try: if seconds < 60: return f"{seconds}秒" elif seconds < 3600: minutes = seconds // 60 remaining_seconds = seconds % 60 if remaining_seconds == 0: return f"{minutes}分钟" else: return f"{minutes}分{remaining_seconds}秒" else: hours = seconds // 3600 remaining_minutes = (seconds % 3600) // 60 if remaining_minutes == 0: return f"{hours}小时" else: return f"{hours}小时{remaining_minutes}分钟" except: return f"{seconds}秒" # 创建全局实例 export_service = ExportService()
294coder/Efficient-MIF
1,319
Pansharpening_Hyper_SR_Matlab_Test_Package/printAllImagesImWriteFR.m
MatrixPrint(:,:,:,1) = I_MS; MatrixPrint(:,:,:,2) = I_BT_H; MatrixPrint(:,:,:,3) = I_BDSD; MatrixPrint(:,:,:,4) = I_C_BDSD; MatrixPrint(:,:,:,5) = I_BDSD_PC; MatrixPrint(:,:,:,6) = I_GS; MatrixPrint(:,:,:,7) = I_GSA; MatrixPrint(:,:,:,8) = I_C_GSA; MatrixPrint(:,:,:,9) = I_PRACS; MatrixPrint(:,:,:,10) = I_AWLP; MatrixPrint(:,:,:,11) = I_MTF_GLP; MatrixPrint(:,:,:,12) = I_MTF_GLP_FS; MatrixPrint(:,:,:,13) = I_MTF_GLP_HPM; MatrixPrint(:,:,:,14) = I_MTF_GLP_HPM_H; MatrixPrint(:,:,:,15) = I_MTF_GLP_HPM_R; MatrixPrint(:,:,:,16) = I_MTF_GLP_CBD; MatrixPrint(:,:,:,17) = I_C_MTF_GLP_CBD; MatrixPrint(:,:,:,18) = I_MF; MatrixPrint(:,:,:,19) = I_FE_HPM; MatrixPrint(:,:,:,20) = I_SR_D; MatrixPrint(:,:,:,21) = I_PWMBF; MatrixPrint(:,:,:,22) = I_TV; MatrixPrint(:,:,:,23) = I_RR; MatrixPrint(:,:,:,24) = I_PNN; MatrixPrint(:,:,:,25) = I_PNN_IDX; MatrixPrint(:,:,:,26) = I_A_PNN; MatrixPrint(:,:,:,27) = I_A_PNN_FT; if size(I_MS,3) == 4 vect_index_RGB = [3,2,1]; else vect_index_RGB = [5,3,2]; end titleImages = algorithms; addpath([pwd,'\Tools']); figure, MP = showImagesAll(MatrixPrint,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0); cd 'Outputs' for ii = 1 : size(MP,4) imwrite(MP(:,:,:,ii),sprintf('%s.png',algorithms{ii})); end imwrite(showPan(I_PAN,0,1,flag_cut_bounds,dim_cut),'PAN.png') cd ..
294coder/Efficient-MIF
2,675
Pansharpening_Hyper_SR_Matlab_Test_Package/ssim_index.m
function [mssim, ssim_map] = ssim_index(img1, img2, K, window, L) %======================================================================== if (nargin < 2 || nargin > 5) mssim = -Inf; ssim_map = -Inf; return; end if (size(img1) ~= size(img2)) mssim = -Inf; ssim_map = -Inf; return; end [M N] = size(img1); if (nargin == 2) if ((M < 11) || (N < 11)) mssim = -Inf; ssim_map = -Inf; return end window = fspecial('gaussian', 11, 1.5); % K(1) = 0.01; % default settings K(2) = 0.03; % L = 255; % end if (nargin == 3) if ((M < 11) || (N < 11)) mssim = -Inf; ssim_map = -Inf; return end window = fspecial('gaussian', 11, 1.5); L = 255; if (length(K) == 2) if (K(1) < 0 || K(2) < 0) mssim = -Inf; ssim_map = -Inf; return; end else mssim = -Inf; ssim_map = -Inf; return; end end if (nargin == 4) [H W] = size(window); if ((H*W) < 4 || (H > M) || (W > N)) mssim = -Inf; ssim_map = -Inf; return end L = 255; if (length(K) == 2) if (K(1) < 0 || K(2) < 0) mssim = -Inf; ssim_map = -Inf; return; end else mssim = -Inf; ssim_map = -Inf; return; end end if (nargin == 5) [H W] = size(window); if ((H*W) < 4 || (H > M) || (W > N)) mssim = -Inf; ssim_map = -Inf; return end if (length(K) == 2) if (K(1) < 0 || K(2) < 0) mssim = -Inf; ssim_map = -Inf; return; end else mssim = -Inf; ssim_map = -Inf; return; end end C1 = (K(1)*L)^2; C2 = (K(2)*L)^2; window = window/sum(sum(window)); img1 = double(img1); img2 = double(img2); mu1 = filter2(window, img1, 'valid'); mu2 = filter2(window, img2, 'valid'); mu1_sq = mu1.*mu1; mu2_sq = mu2.*mu2; mu1_mu2 = mu1.*mu2; sigma1_sq = filter2(window, img1.*img1, 'valid') - mu1_sq; sigma2_sq = filter2(window, img2.*img2, 'valid') - mu2_sq; sigma12 = filter2(window, img1.*img2, 'valid') - mu1_mu2; if (C1 > 0 && C2 > 0) ssim_map = ((2*mu1_mu2 + C1).*(2*sigma12 + C2))./((mu1_sq + mu2_sq + C1).*(sigma1_sq + sigma2_sq + C2)); else numerator1 = 2*mu1_mu2 + C1; numerator2 = 2*sigma12 + C2; denominator1 = mu1_sq + mu2_sq + C1; denominator2 = sigma1_sq + sigma2_sq + C2; ssim_map = ones(size(mu1)); index = (denominator1.*denominator2 > 0); ssim_map(index) = (numerator1(index).*numerator2(index))./(denominator1(index).*denominator2(index)); index = (denominator1 ~= 0) & (denominator2 == 0); ssim_map(index) = numerator1(index)./denominator1(index); end mssim = mean2(ssim_map); return
294coder/Efficient-MIF
1,613
Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_ref_batched_images.m
function res = analysis_ref_batched_images(path, ratio, full_res, const) % data should be [0, max_range] % e.g. wv3: max_range: 2047 data = load(path); gt = data.gt; sr = data.sr; addpath('./Tools') addpath('./Quality_Indices/') Q_block_size = 32; thvalues = 0; L = 11; flag_cut_bounds = 0; if full_res dim_cut = 21; else dim_cut = 30; end bs = size(sr, 1); res = {}; sam = zeros(1, bs); ergas = zeros(1, bs); scc = zeros(1, bs); qn = zeros(1, bs); qave = zeros(1, bs); psnr = zeros(1, bs); ssim = zeros(1, bs); for i = (1:bs) sr1 = permute(squeeze(sr(i, :, :, :)), [2, 3, 1]); gt1 = permute(squeeze(gt(i, :, :, :)), [2, 3, 1]); [q_avg_gt, sam_gt, ergas_gt, scc_gt, q_gt] = indexes_evaluation(sr1, gt1, ratio, L, Q_block_size, flag_cut_bounds, dim_cut, thvalues); [psnr_gt, ssim_gt] = quality_assess(sr1 / const, gt1 / const); sam(i) = sam_gt; ergas(i) = ergas_gt; scc(i) = scc_gt; qn(i) = q_gt; qave(i) = q_avg_gt; psnr(i) = psnr_gt; ssim(i) = ssim_gt; fprintf("sample %d - sam: %f, ergas: %f, scc: %f, qn: %f, q_ave: %f, psnr: %f, ssim: %f \n", i, sam_gt, ergas_gt, scc_gt, q_gt, q_avg_gt, psnr_gt, ssim_gt) end res.sam = [mean(sam), std(sam)]; res.ergas = [mean(ergas), std(ergas)]; res.scc = [mean(scc), std(scc)]; res.qn = [mean(qn), std(qn)]; res.qave = [mean(qave), std(qave)]; res.psnr = [mean(psnr), std(psnr)]; res.ssim = [mean(ssim), std(ssim)]; end
281677160/openwrt-package
6,592
luci-app-passwall/root/usr/share/passwall/0_default_config
config global option enabled '0' option socks_enabled '0' option tcp_node_socks_port '1070' option filter_proxy_ipv6 '1' option dns_shunt 'chinadns-ng' option dns_mode 'tcp' option remote_dns '1.1.1.1' list smartdns_remote_dns 'https://1.1.1.1/dns-query' option use_default_dns 'direct' option chinadns_ng_default_tag 'none' option dns_redirect '1' option use_direct_list '1' option use_proxy_list '1' option use_block_list '1' option use_gfw_list '1' option chn_list 'direct' option tcp_proxy_mode 'proxy' option udp_proxy_mode 'proxy' option localhost_proxy '1' option client_proxy '1' option acl_enable '0' option log_tcp '0' option log_udp '0' option loglevel 'error' option trojan_loglevel '4' option log_chinadns_ng '0' config global_haproxy option balancing_enable '0' config global_delay option start_daemon '1' option start_delay '60' config global_forwarding option tcp_no_redir_ports 'disable' option udp_no_redir_ports 'disable' option tcp_proxy_drop_ports 'disable' option udp_proxy_drop_ports '443' option tcp_redir_ports '22,25,53,80,143,443,465,587,853,873,993,995,5222,8080,8443,9418' option udp_redir_ports '1:65535' option accept_icmp '0' option use_nft '0' option tcp_proxy_way 'redirect' option ipv6_tproxy '0' config global_xray option sniffing_override_dest '0' config global_singbox option sniff_override_destination '0' config global_other option auto_detection_time 'tcping' option show_node_info '0' config global_rules option auto_update '0' option chnlist_update '1' option chnroute_update '1' option chnroute6_update '1' option gfwlist_update '1' option geosite_update '0' option geoip_update '0' list gfwlist_url 'https://fastly.jsdelivr.net/gh/Loyalsoldier/v2ray-rules-dat@release/gfw.txt' list chnroute_url 'https://ispip.clang.cn/all_cn.txt' list chnroute_url 'https://fastly.jsdelivr.net/gh/gaoyifan/china-operator-ip@ip-lists/china.txt' list chnroute6_url 'https://ispip.clang.cn/all_cn_ipv6.txt' list chnroute6_url 'https://fastly.jsdelivr.net/gh/gaoyifan/china-operator-ip@ip-lists/china6.txt' list chnlist_url 'https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/accelerated-domains.china.conf' list chnlist_url 'https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/apple.china.conf' option v2ray_location_asset '/usr/share/v2ray/' option geoip_url 'https://github.com/Loyalsoldier/geoip/releases/latest/download/geoip.dat' option geosite_url 'https://github.com/Loyalsoldier/v2ray-rules-dat/releases/latest/download/geosite.dat' config global_app option sing_box_file '/usr/bin/sing-box' option xray_file '/usr/bin/xray' option hysteria_file '/usr/bin/hysteria' config global_subscribe option filter_keyword_mode '1' list filter_discard_list '距离下次重置剩余' list filter_discard_list '套餐到期' list filter_discard_list '过期时间' list filter_discard_list '剩余流量' list filter_discard_list 'QQ群' list filter_discard_list '官网' config nodes 'myshunt' option remarks '分流总节点' option type 'Xray' option protocol '_shunt' option DirectGame '_direct' option ProxyGame '_default' option AIGC '_default' option Streaming '_default' option Proxy '_default' option Direct '_direct' option default_node '_direct' option domainStrategy 'IPOnDemand' config shunt_rules 'DirectGame' option remarks 'DirectGame' option domain_list '# steam直连域名获取国内CDN走国内线路下载 cm.steampowered.com steamserver.net # steam国内CDN华为云 steampipe.steamcontent.tnkjmec.com # steam国内CDN白山云 st.dl.eccdnx.com st.dl.bscstorage.net st.dl.pinyuncloud.com # steam国内CDN新流云(原金山云)(支持ipv6) dl.steam.clngaa.com # steam国内CDN网宿 cdn.mileweb.cs.steampowered.com.8686c.com cdn-ws.content.steamchina.com # steam国内CDN腾讯云 (蒸汽中国独占) cdn-qc.content.steamchina.com # steam国内CDN阿里云(支持ipv6) cdn-ali.content.steamchina.com xz.pphimalayanrt.com lv.queniujq.cn alibaba.cdn.steampipe.steamcontent.com # 国内游戏geosite域名 geosite:category-games@cn' option ip_list '# steam直连IP 45.121.184.0/24 103.10.124.0/23 103.28.54.0/24 146.66.152.0/24 146.66.155.0/24 153.254.86.0/24 155.133.224.0/22 155.133.230.0/24 155.133.232.0/23 155.133.234.0/24 155.133.236.0/22 155.133.240.0/23 155.133.244.0/23 155.133.246.0/24 155.133.248.0/21 162.254.192.0/21 185.25.182.0/23 190.217.32.0/22 192.69.96.0/22 205.196.6.0/24 208.64.200.0/22 208.78.164.0/22 205.185.194.0/24' config shunt_rules 'ProxyGame' option remarks 'ProxyGame' option domain_list '# steam 商店/客服/聊天/网页布局/API/二维码/Google云同步 代理URL steamcommunity.com www.steamcommunity.com store.steampowered.com checkout.steampowered.com api.steampowered.com help.steampowered.com login.steampowered.com store.akamai.steamstatic.com steambroadcast.akamaized.net steamvideo-a.akamaihd.net steamusercontent-a.akamaihd.net steamstore-a.akamaihd.net steamcommunity-a.akamaihd.net steamcdn-a.akamaihd.net steamuserimages-a.akamaihd.net community.akamai.steamstatic.com avatars.akamai.steamstatic.com community.steamstatic.com cdn.akamai.steamstatic.com avatars.steamstatic.com shared.akamai.steamstatic.com clan.akamai.steamstatic.com cdn.cloudflare.steamstatic.com community.cloudflare.steamstatic.com store.cloudflare.steamstatic.com avatars.cloudflare.steamstatic.com clan.cloudflare.steamstatic.com shared.cloudflare.steamstatic.com steam-chat.com steamcloud-ugc.storage.googleapis.com steamcloud-eu-ams.storage.googleapis.com steamcloud-eu-fra.storage.googleapis.com steamcloud-finland.storage.googleapis.com steamcloud-saopaulo.storage.googleapis.com steamcloud-singapore.storage.googleapis.com steamcloud-sydney.storage.googleapis.com steamcloud-taiwan.storage.googleapis.com steamcloud-eu.storage.googleapis.com geosite:category-games' config shunt_rules 'AIGC' option remarks 'AIGC' option domain_list 'geosite:category-ai-!cn geosite:apple-intelligence' config shunt_rules 'Streaming' option remarks 'Streaming' option domain_list 'geosite:netflix geosite:disney' config shunt_rules 'Proxy' option remarks 'Proxy' option domain_list 'geosite:geolocation-!cn' option ip_list '149.154.160.0/20 91.108.4.0/22 91.108.56.0/24 109.239.140.0/24 67.198.55.0/24 8.8.4.4 8.8.8.8 208.67.222.222 208.67.220.220 1.1.1.1 1.1.1.2 1.0.0.1 9.9.9.9 149.112.112.112 2001:67c:4e8::/48 2001:b28:f23c::/48 2001:b28:f23d::/48 2001:b28:f23f::/48 2001:b28:f242::/48 2001:4860:4860::8888 2001:4860:4860::8844 2606:4700:4700::1111 2606:4700:4700::1001' config shunt_rules 'Direct' option remarks 'Direct' option domain_list 'geosite:cn' option ip_list '223.5.5.5/32 223.6.6.6/32 119.29.29.29/32 180.76.76.76/32 114.114.114.114/32 114.114.115.115/32 1.12.12.12/32 120.53.53.53/32 geoip:cn geoip:private'
2977094657/DidaAPI
16,185
services/wechat_service.py
"""微信登录服务模块""" import re import uuid from typing import Optional, Dict, Any, Tuple import httpx from utils import app_logger from core import config, db, urls from models import WeChatQRResponse, WeChatValidateResponse, PasswordLoginRequest class WeChatLoginService: """微信登录服务类""" def __init__(self): self.request_config = config.get('request_config', {}) self.client = httpx.AsyncClient(timeout=self.request_config.get('timeout', 30.0)) async def get_qr_code(self, state: str = "Lw==") -> Optional[WeChatQRResponse]: """ 获取微信登录二维码 Args: state: 状态参数,默认为 "Lw==" Returns: WeChatQRResponse: 包含二维码URL和密钥的响应对象 """ try: # 使用统一的URL构建函数 qr_url = urls.build_wechat_qr_url(state) app_logger.info(f"请求微信二维码: {qr_url}") # 发送请求 response = await self.client.get(qr_url) response.raise_for_status() # 记录完整响应 app_logger.debug(f"微信二维码响应状态: {response.status_code}") app_logger.debug(f"微信二维码响应头: {dict(response.headers)}") app_logger.debug(f"微信二维码响应内容长度: {len(response.text)}") # 解析HTML中的二维码图片链接 qr_code_key = self._extract_qr_code_key(response.text) if not qr_code_key: app_logger.error("未能从响应中提取二维码密钥") return None # 构建完整的二维码图片URL qr_code_url = f"{urls.WECHAT_URLS['qr_image_base_url']}/{qr_code_key}" # 记录到数据库 db.log_wechat_login(qr_code_key=qr_code_key, state=state) app_logger.info(f"成功获取二维码: {qr_code_url}") return WeChatQRResponse( qr_code_url=qr_code_url, qr_code_key=qr_code_key, state=state ) except Exception as e: app_logger.error(f"获取微信二维码失败: {e}") return None def _extract_qr_code_key(self, html_content: str) -> Optional[str]: """ 从HTML内容中提取二维码密钥 Args: html_content: HTML响应内容 Returns: str: 16位二维码密钥,如果未找到则返回None """ try: # 查找二维码图片标签 pattern = r'<img[^>]*class="[^"]*qrcode[^"]*"[^>]*src="([^"]*)"' match = re.search(pattern, html_content) if match: src_url = match.group(1) app_logger.debug(f"找到二维码图片src: {src_url}") # 提取最后16位字符 qr_code_key = src_url.split('/')[-1] if len(qr_code_key) >= 16: qr_code_key = qr_code_key[-16:] app_logger.info(f"提取到二维码密钥: {qr_code_key}") return qr_code_key else: app_logger.warning(f"二维码密钥长度不足16位: {qr_code_key}") # 如果上面的方法失败,尝试其他模式 patterns = [ r'/connect/qrcode/([a-zA-Z0-9]{16})', r'qrcode/([a-zA-Z0-9]{16})', r'src="[^"]*?([a-zA-Z0-9]{16})"' ] for pattern in patterns: matches = re.findall(pattern, html_content) if matches: qr_code_key = matches[0] app_logger.info(f"通过备用模式提取到二维码密钥: {qr_code_key}") return qr_code_key app_logger.error("未能从HTML中提取二维码密钥") app_logger.debug(f"HTML内容片段: {html_content[:500]}...") return None except Exception as e: app_logger.error(f"提取二维码密钥时发生错误: {e}") return None async def poll_qr_status(self, qr_code_key: str, max_attempts: int = 60) -> Optional[WeChatValidateResponse]: """ 轮询二维码状态,检查是否已扫码登录 Args: qr_code_key: 二维码密钥 max_attempts: 最大轮询次数,默认60次(约5分钟) Returns: WeChatValidateResponse: 登录结果 """ import asyncio for attempt in range(max_attempts): try: app_logger.info(f"轮询二维码状态,第 {attempt + 1}/{max_attempts} 次") # 使用统一的URL构建函数 poll_url = urls.build_wechat_poll_url(qr_code_key) headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36', 'Referer': 'https://open.weixin.qq.com/', 'Accept': '*/*', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8' } response = await self.client.get(poll_url, headers=headers) response_text = response.text app_logger.debug(f"轮询响应: {response_text}") # 解析微信轮询响应 # 实际响应格式是: window.wx_errcode=405;window.wx_code='xxx'; if "window.wx_errcode" in response_text: import re # 提取错误码 errcode_match = re.search(r'window\.wx_errcode\s*=\s*(\d+)', response_text) code_match = re.search(r"window\.wx_code\s*=\s*'([^']*)'", response_text) if errcode_match: errcode = int(errcode_match.group(1)) wx_code = code_match.group(1) if code_match else '' app_logger.info(f"轮询状态 - errcode: {errcode}, wx_code: {wx_code}") if errcode == 405 and wx_code: # 登录成功,获得了授权码 app_logger.info(f"检测到登录成功,获得授权码: {wx_code}") # 使用获得的code进行验证 return await self.validate_wechat_login(wx_code) elif errcode == 404: app_logger.info("等待扫码...") elif errcode == 403: app_logger.info("二维码已扫描,等待用户确认") elif errcode == 408: app_logger.info("二维码已过期") break elif errcode == 400: app_logger.info("二维码已失效") break # 等待5秒后继续轮询 await asyncio.sleep(5) except Exception as e: app_logger.error(f"轮询二维码状态失败: {e}") await asyncio.sleep(5) app_logger.warning("轮询超时,未检测到登录") return WeChatValidateResponse( success=False, message="轮询超时,请重新获取二维码", token=None, user_info=None, cookies=None, raw_response={"error": "polling_timeout"} ) async def validate_wechat_login(self, code: str, state: str = "Lw==") -> Optional[WeChatValidateResponse]: """ 验证微信登录 Args: code: 扫码后获得的验证码 state: 状态参数 Returns: WeChatValidateResponse: 验证响应对象 """ try: # 使用统一的URL构建函数 validate_url = urls.build_wechat_validate_url(code, state) app_logger.info(f"验证微信登录: {validate_url}") # 设置请求头,模拟浏览器请求 headers = { 'Accept': '*/*', 'Accept-Encoding': 'gzip, deflate, br, zstd', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7', 'Content-Type': 'application/json', 'Origin': 'https://dida365.com', 'Referer': 'https://dida365.com/', 'Sec-Ch-Ua': '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"', 'Sec-Ch-Ua-Mobile': '?0', 'Sec-Ch-Ua-Platform': '"Windows"', 'Sec-Fetch-Dest': 'empty', 'Sec-Fetch-Mode': 'cors', 'Sec-Fetch-Site': 'same-site', 'User-Agent': self.request_config.get('user_agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'), 'X-Device': self.request_config.get('device_info', '{}') } # 发送验证请求 response = await self.client.get(validate_url, headers=headers) # 记录详细的响应信息 app_logger.info(f"验证响应状态码: {response.status_code}") app_logger.info(f"验证响应头: {dict(response.headers)}") # 提取cookies cookies = {} if hasattr(response, 'cookies') and response.cookies: for cookie_name, cookie_value in response.cookies.items(): cookies[cookie_name] = cookie_value # 同时从Set-Cookie头中解析cookies set_cookie_header = response.headers.get('set-cookie', '') if set_cookie_header: import re # 解析Set-Cookie头 cookie_matches = re.findall(r'([^=]+)=([^;]+)', set_cookie_header) for name, value in cookie_matches: cookies[name.strip()] = value.strip() app_logger.info(f"响应cookies: {cookies}") # 尝试解析JSON响应 response_data = {} try: response_data = response.json() app_logger.info(f"验证响应JSON: {response_data}") except Exception as json_error: app_logger.warning(f"响应不是有效的JSON: {json_error}") app_logger.info(f"响应文本内容: {response.text}") response_data = {"raw_text": response.text} # 检查是否成功 success = response.status_code == 200 # 提取认证令牌 token = cookies.get('t', '') csrf_token = cookies.get('_csrf_token', '') # 保存会话信息 if success and token: session_id = str(uuid.uuid4()) session_data = { 'session_id': session_id, 'token': token, 'csrf_token': csrf_token, 'cookies': cookies, 'is_active': True } db.save_user_session(session_data) # 自动设置滴答清单API认证会话 try: from services.dida_service import dida_service dida_service.set_auth_session(token, csrf_token) app_logger.info("已自动设置滴答清单API认证会话") except Exception as e: app_logger.warning(f"自动设置滴答清单API认证会话失败: {e}") # 记录登录日志 db.log_wechat_login( qr_code_key="", # 这里可能需要从之前的记录中关联 validation_code=code, state=state, response_data={ 'status_code': response.status_code, 'headers': dict(response.headers), 'cookies': cookies, 'json_data': response_data }, status='success' if success else 'failed' ) return WeChatValidateResponse( success=success, message="登录成功" if success else "登录失败", token=token if token else None, user_info=response_data.get('user', {}), cookies=cookies, raw_response=response_data ) except Exception as e: app_logger.error(f"验证微信登录失败: {e}") # 记录失败日志 db.log_wechat_login( qr_code_key="", validation_code=code, state=state, response_data={'error': str(e)}, status='failed' ) return WeChatValidateResponse( success=False, message=f"验证失败: {str(e)}", token=None, user_info=None, cookies=None, raw_response={'error': str(e)} ) async def password_login(self, username: str, password: str) -> dict: """ 密码登录滴答清单 Args: username: 登录账户(邮箱或手机号) password: 登录密码 Returns: dict: 原始响应数据 """ try: # 使用统一的URL构建函数 login_url = urls.build_password_login_url(wc=True, remember=True) app_logger.info(f"密码登录请求: {login_url}") # 构建请求体 login_data = { "username": username, "password": password } # 设置请求头,模拟浏览器请求 headers = { 'Accept': 'application/json, text/plain, */*', 'Accept-Encoding': 'gzip, deflate, br, zstd', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7', 'Content-Type': 'application/json', 'Origin': 'https://dida365.com', 'Referer': 'https://dida365.com/', 'Sec-Ch-Ua': '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"', 'Sec-Ch-Ua-Mobile': '?0', 'Sec-Ch-Ua-Platform': '"Windows"', 'Sec-Fetch-Dest': 'empty', 'Sec-Fetch-Mode': 'cors', 'Sec-Fetch-Site': 'same-site', 'User-Agent': self.request_config.get('user_agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'), 'X-Device': self.request_config.get('device_info', '{}') } # 发送POST请求 response = await self.client.post(login_url, json=login_data, headers=headers) # 记录详细的响应信息 app_logger.info(f"密码登录响应状态码: {response.status_code}") app_logger.info(f"密码登录响应头: {dict(response.headers)}") # 提取cookies cookies = {} if hasattr(response, 'cookies') and response.cookies: for cookie_name, cookie_value in response.cookies.items(): cookies[cookie_name] = cookie_value # 同时从Set-Cookie头中解析cookies set_cookie_header = response.headers.get('set-cookie', '') if set_cookie_header: import re # 解析Set-Cookie头 cookie_matches = re.findall(r'([^=]+)=([^;]+)', set_cookie_header) for name, value in cookie_matches: cookies[name.strip()] = value.strip() app_logger.info(f"密码登录响应cookies: {cookies}") # 尝试解析JSON响应 response_data = {} try: response_data = response.json() app_logger.info(f"密码登录响应JSON: {response_data}") except Exception as json_error: app_logger.warning(f"响应不是有效的JSON: {json_error}") app_logger.info(f"响应文本内容: {response.text}") response_data = {"raw_text": response.text} # 检查是否成功 success = response.status_code == 200 and 'token' in response_data # 如果成功,保存会话信息 if success: token = response_data.get('token', '') if token: session_id = str(uuid.uuid4()) session_data = { 'session_id': session_id, 'token': token, 'csrf_token': '', # 密码登录可能不返回CSRF token 'cookies': cookies, 'is_active': True } db.save_user_session(session_data) # 自动设置滴答清单API认证会话 try: from services.dida_service import dida_service dida_service.set_auth_session(token, '') app_logger.info("已自动设置滴答清单API认证会话") except Exception as e: app_logger.warning(f"自动设置滴答清单API认证会话失败: {e}") # 直接返回原始响应 return response_data except Exception as e: app_logger.error(f"密码登录失败: {e}") # 返回错误响应 return {'error': str(e)} async def close(self): """关闭HTTP客户端""" await self.client.aclose() # 全局微信登录服务实例 wechat_service = WeChatLoginService()
2977094657/DidaAPI
5,269
services/habit_service.py
"""习惯管理服务模块""" import httpx from typing import Optional from utils import app_logger from core import urls # 不再使用响应模型,直接返回原始响应 class HabitService: """习惯管理服务类""" def __init__(self): self.client = httpx.AsyncClient(timeout=30.0) def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict: """构建认证请求头""" return { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36', 'Accept': 'application/json, text/plain, */*', 'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8', 'Content-Type': 'application/json', 'X-Requested-With': 'XMLHttpRequest', 'X-Tz': 'Asia/Shanghai', } def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict: """构建认证cookies""" return { 't': auth_token, '_csrf_token': csrf_token } async def get_habits(self, auth_token: str, csrf_token: str) -> dict: """ 获取习惯列表 Args: auth_token: 认证令牌 csrf_token: CSRF令牌 Returns: dict: 原始响应数据 """ try: url = urls.build_dida_api_url(urls.DIDA_HABIT_APIS["get_habits"]) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) app_logger.info(f"请求获取习惯列表: {url}") response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: response_data = response.json() app_logger.info(f"成功获取习惯列表,习惯数: {len(response_data) if isinstance(response_data, list) else 0}") # 直接返回原始响应 return response_data else: app_logger.error(f"获取习惯列表失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"获取习惯列表时发生错误: {e}") return {"error": str(e)} async def get_week_current_statistics(self, auth_token: str, csrf_token: str) -> dict: """ 获取本周习惯打卡统计 Args: auth_token: 认证令牌 csrf_token: CSRF令牌 Returns: dict: 原始响应数据 """ try: url = urls.build_dida_api_url(urls.DIDA_HABIT_APIS["week_current_statistics"]) headers = self._build_auth_headers(auth_token, csrf_token) cookies = self._build_auth_cookies(auth_token, csrf_token) app_logger.info(f"请求获取本周习惯打卡统计: {url}") response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: response_data = response.json() app_logger.info("成功获取本周习惯打卡统计") # 直接返回原始响应 return response_data else: app_logger.error(f"获取本周习惯打卡统计失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"获取本周习惯打卡统计时发生错误: {e}") return {"error": str(e)} async def export_habits(self, auth_token: str, csrf_token: str) -> dict: """ 导出习惯数据(Excel格式) Args: auth_token: 认证令牌 csrf_token: CSRF令牌 Returns: dict: 包含文件内容和元数据的响应 """ try: url = urls.build_dida_api_url(urls.DIDA_HABIT_APIS["export_habits"]) # 对于文件下载,需要修改Accept头 headers = self._build_auth_headers(auth_token, csrf_token) headers['Accept'] = '*/*' headers['X-CSRFToken'] = csrf_token cookies = self._build_auth_cookies(auth_token, csrf_token) app_logger.info(f"请求导出习惯数据: {url}") response = await self.client.get(url, headers=headers, cookies=cookies) if response.status_code == 200: # 获取文件名 content_disposition = response.headers.get('content-disposition', '') filename = 'habits_export.xlsx' if 'filename=' in content_disposition: filename = content_disposition.split('filename=')[1].split(';')[0].strip('"') app_logger.info(f"成功导出习惯数据,文件名: {filename}") # 返回文件内容和元数据 return { "filename": filename, "content_type": response.headers.get('content-type', 'application/vnd.openxmlformats-officedocument.spreadsheetml.sheet'), "content": response.content, "size": len(response.content) } else: app_logger.error(f"导出习惯数据失败,状态码: {response.status_code}") return {"error": f"HTTP {response.status_code}", "text": response.text} except Exception as e: app_logger.error(f"导出习惯数据时发生错误: {e}") return {"error": str(e)} async def close(self): """关闭HTTP客户端""" await self.client.aclose() # 全局习惯服务实例 habit_service = HabitService()
294coder/Efficient-MIF
7,005
Pansharpening_Hyper_SR_Matlab_Test_Package/rantest.m
%This is a demo to run fusion algorithms on Reduced Resolution % LJ Deng(UESTC) TJ Zhang % 2022-05-07 clear; close all; %% =======load directors======== % Tools addpath([pwd,'/Tools']); % Select algorithms to run algorithms = {'GT','EXP','BT-H','BDSD-PC','C-GSA','SR-D',... 'MTF-GLP-HPM-R','MTF-GLP-FS','TV','PanNet','PNN','DiCNN','FusionNet','LAGConv','MSDCNN','BDPN'};%'PNN' location1 = [2 40 4 43]; %default: data6: [10 50 1 60]; data7:[140 180 5 60] location2 = []; %default: data6: [190 240 5 60]; data7:[190 235 120 150] sensor = 'WV3'; %% =======read Multiple TestData_wv3.h5 (four 512x512 WV3 simulated data)======== file_test = '/Data2/DataSet/pansharpening_2/test_data/WV3/test_wv3_multiExm1.h5'; gt_multiExm_tmp = h5read(file_test,'/gt'); % WxHxCxN=1x2x3x4 gt_multiExm = permute(gt_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 ms_multiExm_tmp = h5read(file_test,'/ms'); % WxHxCxN=1x2x3x4 ms_multiExm = permute(ms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 lms_multiExm_tmp = h5read(file_test,'/lms'); % WxHxCxN=1x2x3x4 lms_multiExm = permute(lms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 pan_multiExm_tmp = h5read(file_test,'/pan'); % WxHxCxN=1x2x3x4 pan_multiExm = permute(pan_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 data_name = '3_EPS/WV3/multi/'; % director to save EPS figures %% ==========Read each Data==================== exm_num = size(ms_multiExm, 1); for i = 1 : 20% i = 1 or 2 ... %% read each data HRMS_tmp = gt_multiExm(i, :, :, :); % I_GT I_GT = squeeze(HRMS_tmp); LRMS_tmp = ms_multiExm(i, :, :, :); % I_MS_LR I_MS_LR = squeeze(LRMS_tmp); LMS_tmp = lms_multiExm(i, :, :, :); % I_MS I_MS = squeeze(LMS_tmp); PAN_tmp = pan_multiExm(i, :, :, :); % I_PAN I_PAN = squeeze(PAN_tmp); %% Initialization of the Matrix of Results NumIndexes = 5; MatrixResults = zeros(numel(algorithms),NumIndexes); alg = 0; %% load Indexes for WV3_RR sensor = 'WV3'; Qblocks_size = 32; bicubic = 0;% Interpolator flag_cut_bounds = 0;% Cut Final Image dim_cut = 21;% Cut Final Image thvalues = 0;% Threshold values out of dynamic range printEPS = 0;% Print Eps ratio = 4;% Resize Factor L = 11;% Radiometric Resolution %% show I_MS_LR, I_GT, PAN Imgs: showImage8_zoomin(I_MS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); showPan(I_PAN,printEPS,2,flag_cut_bounds,dim_cut); print('-deps', strcat(data_name, num2str(i-1), '_pan', '.eps')) %% ======GT =================== if ismember('GT',algorithms) alg = alg + 1; [Q_avg_GT, SAM_GT, ERGAS_GT, SCC_GT, Q_GT] = indexes_evaluation(I_GT,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_GT,Q_avg_GT,SAM_GT,ERGAS_GT,SCC_GT]; MatrixImage(:,:,:,alg) = I_GT; Q_avg_GT_multiexm(i) = Q_avg_GT; SAM_GT_multiexm(i) = SAM_GT; ERGAS_GT_multiexm(i) = ERGAS_GT; SCC_GT_multiexm(i) = SCC_GT; Q_GT_multiexm(i) = Q_GT; showImage8_zoomin(I_GT,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); print('-depsc', strcat(data_name, num2str(i-1), '_gt', '.eps')) end %% ====== 4) BDPN Method ====== file_bdpn = 'bdpn_wv3_rs'; load(strcat('/Data2/DataSet/pansharpening_2/results/p2/wv3_multiExm1.h5/BDPN/Test/model_2022-05-21-10-46-03/results/output_mulExm_' , num2str(i-1), '.mat')) % load i-th image for DiCNN I_bdpn = double(sr); if ismember('BDPN',algorithms) alg = alg + 1; [Q_avg_bdpn, SAM_bdpn, ERGAS_bdpn, SCC_bdpn, Q_bdpn] = indexes_evaluation(I_bdpn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_bdpn,Q_avg_bdpn,SAM_bdpn,ERGAS_bdpn,SCC_bdpn]; MatrixImage(:,:,:,alg) = I_bdpn; Q_avg_bdpn_multiexm(i) = Q_avg_bdpn; SAM_bdpn_multiexm(i) = SAM_bdpn; ERGAS_bdpn_multiexm(i) = ERGAS_bdpn; SCC_bdpn_multiexm(i) = SCC_bdpn; Q_bdpn_multiexm(i) = Q_bdpn; showImage8_zoomin(I_bdpn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); print('-depsc', strcat(data_name, num2str(i-1),'_bdpn.eps')) end % % % %% ====== 7) APNN Method ====== % file_apnn = 'apnn_wv3_rs'; % load(strcat('2_DL_Result/WV3/APNN/', file_apnn, num2str(i-1), '.mat')) % load i-th image for DiCNN % I_apnn = 2047*double(apnn_wv3_rs); % % if ismember('APNN',algorithms) % alg = alg + 1; % [Q_avg_apnn, SAM_apnn, ERGAS_apnn, SCC_apnn, Q_apnn] = indexes_evaluation(I_apnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_apnn,Q_avg_apnn,SAM_apnn,ERGAS_apnn,SCC_apnn]; % MatrixImage(:,:,:,alg) = I_apnn; % % Q_avg_apnn_multiexm(i) = Q_avg_apnn; % SAM_apnn_multiexm(i) = SAM_apnn; % ERGAS_apnn_multiexm(i) = ERGAS_apnn; % SCC_apnn_multiexm(i) = SCC_apnn; % Q_apnn_multiexm(i) = Q_apnn; % % showImage8_zoomin(I_apnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % print('-depsc', strcat(data_name, num2str(i-1),'_apnn.eps')) % end end %% Print in LATEX %% View All if size(I_GT,3) == 4 vect_index_RGB = [3,2,1]; else vect_index_RGB = [5,3,2]; end titleImages = algorithms; figure, showImagesAll(MatrixImage,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0); %% ======Display the final average performance ======= % GT: average Q_avg % bdpn: average Q_avg avg_Q_bdpn_multiexm = mean(Q_bdpn_multiexm); std_Q_bdpn_multiexm = std(Q_bdpn_multiexm); avg_Q_avg_bdpn_multiexm = mean(Q_avg_bdpn_multiexm); std_Q_avg_bdpn_multiexm = std(Q_avg_bdpn_multiexm); avg_SAM_bdpn_multiexm = mean(SAM_bdpn_multiexm); std_SAM_bdpn_multiexm = std(SAM_bdpn_multiexm); avg_ERGAS_bdpn_multiexm = mean(ERGAS_bdpn_multiexm); std_ERGAS_bdpn_multiexm = std(ERGAS_bdpn_multiexm); avg_SCC_bdpn_multiexm = mean(SCC_bdpn_multiexm); std_SCC_bdpn_multiexm = std(SCC_bdpn_multiexm); Avg_MatrixResults(17,:) = [avg_Q_bdpn_multiexm, std_Q_bdpn_multiexm, avg_Q_avg_bdpn_multiexm, std_Q_avg_bdpn_multiexm, ... avg_SAM_bdpn_multiexm, std_SAM_bdpn_multiexm, avg_ERGAS_bdpn_multiexm, std_ERGAS_bdpn_multiexm,... avg_SCC_bdpn_multiexm, std_SCC_bdpn_multiexm]; fprintf('\n') disp('#######################################################') disp(['Display the performance for:', num2str(1:i)]) disp('#######################################################') disp(' |====Q====|===Q_avg===|=====SAM=====|======ERGAS=======|=======SCC=======') MatrixResults
294coder/Efficient-MIF
1,349
Pansharpening_Hyper_SR_Matlab_Test_Package/printAllImagesImWriteRR.m
MatrixPrint(:,:,:,1) = I_GT; MatrixPrint(:,:,:,2) = I_MS; MatrixPrint(:,:,:,3) = I_BT_H; MatrixPrint(:,:,:,4) = I_BDSD; MatrixPrint(:,:,:,5) = I_C_BDSD; MatrixPrint(:,:,:,6) = I_BDSD_PC; MatrixPrint(:,:,:,7) = I_GS; MatrixPrint(:,:,:,8) = I_GSA; MatrixPrint(:,:,:,9) = I_C_GSA; MatrixPrint(:,:,:,10) = I_PRACS; MatrixPrint(:,:,:,11) = I_AWLP; MatrixPrint(:,:,:,12) = I_MTF_GLP; MatrixPrint(:,:,:,13) = I_MTF_GLP_FS; MatrixPrint(:,:,:,14) = I_MTF_GLP_HPM; MatrixPrint(:,:,:,15) = I_MTF_GLP_HPM_H; MatrixPrint(:,:,:,16) = I_MTF_GLP_HPM_R; MatrixPrint(:,:,:,17) = I_MTF_GLP_CBD; MatrixPrint(:,:,:,18) = I_C_MTF_GLP_CBD; MatrixPrint(:,:,:,19) = I_MF; MatrixPrint(:,:,:,20) = I_FE_HPM; MatrixPrint(:,:,:,21) = I_SR_D; MatrixPrint(:,:,:,22) = I_PWMBF; MatrixPrint(:,:,:,23) = I_TV; MatrixPrint(:,:,:,24) = I_RR; MatrixPrint(:,:,:,25) = I_PNN; MatrixPrint(:,:,:,26) = I_PNN_IDX; MatrixPrint(:,:,:,27) = I_A_PNN; MatrixPrint(:,:,:,28) = I_A_PNN_FT; if size(I_MS,3) == 4 vect_index_RGB = [3,2,1]; else vect_index_RGB = [5,3,2]; end titleImages = algorithms; addpath([pwd,'\Tools']); figure, MP = showImagesAll(MatrixPrint,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0); cd 'Outputs' for ii = 1 : size(MP,4) imwrite(MP(:,:,:,ii),sprintf('%s.png',algorithms{ii})); end imwrite(showPan(I_PAN,0,1,flag_cut_bounds,dim_cut),'PAN.png') cd ..
294coder/Efficient-MIF
5,360
Pansharpening_Hyper_SR_Matlab_Test_Package/Avg_RR_Assessment.tex
\begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0001&0.0025&0.0000&0.0016&0.0008&0.0264\\\hline \textbf{BDSD-PC}&0.0001&0.0032&0.0000&0.0015&0.0008&0.0259\\\hline \textbf{MTF-GLP-HPM-R}&0.0001&0.0017&0.0000&0.0010&0.0008&0.0278\\\hline \textbf{MTF-GLP-FS}&0.0000&0.0016&0.0000&0.0012&0.0008&0.0277\\\hline \textbf{TV}&0.0001&0.0018&0.0000&0.0008&0.0008&0.0279\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0001&0.0025&0.0000&0.0016&0.0008&0.0264\\\hline \textbf{BDSD-PC}&0.0001&0.0032&0.0000&0.0015&0.0008&0.0259\\\hline \textbf{MTF-GLP-HPM-R}&0.0001&0.0017&0.0000&0.0010&0.0008&0.0278\\\hline \textbf{MTF-GLP-FS}&0.0000&0.0016&0.0000&0.0012&0.0008&0.0277\\\hline \textbf{TV}&0.0001&0.0018&0.0000&0.0008&0.0008&0.0279\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline \textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline \textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline \textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline \textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline \textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline \textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline \textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline \textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline \textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline \textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline \textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline \textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0938&0.0341&0.1712&0.0388&0.7520&0.0572\\\hline \textbf{BDSD-PC}&0.0960&0.0292&0.1652&0.0357&0.7553&0.0506\\\hline \textbf{MTF-GLP-HPM-R}&0.0367&0.0133&0.1445&0.0341&0.8243&0.0393\\\hline \textbf{MTF-GLP-FS}&0.0373&0.0140&0.1451&0.0339&0.8233&0.0398\\\hline \textbf{TV}&0.1685&0.0480&0.1277&0.0460&0.7269&0.0724\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline \textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline \textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline \textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline \textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.2759&0.1309&0.1835&0.0861&0.5989&0.1528\\\hline \textbf{BDSD-PC}&0.2171&0.0657&0.1929&0.1004&0.6373&0.1228\\\hline \textbf{MTF-GLP-HPM-R}&0.0744&0.0372&0.1606&0.0787&0.7787&0.0918\\\hline \textbf{MTF-GLP-FS}&0.0704&0.0325&0.1762&0.0777&0.7677&0.0908\\\hline \textbf{TV}&0.0820&0.0366&0.1337&0.0835&0.7979&0.1043\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0878&0.0332&0.1712&0.0388&0.7569&0.0564\\\hline \textbf{BDSD-PC}&0.0899&0.0286&0.1636&0.0362&0.7619&0.0505\\\hline \textbf{MTF-GLP-HPM-R}&0.0365&0.0130&0.1574&0.0360&0.8121&0.0404\\\hline \textbf{MTF-GLP-FS}&0.0371&0.0137&0.1583&0.0358&0.8107&0.0408\\\hline \textbf{TV}&0.0696&0.0300&0.1448&0.0378&0.7957&0.0443\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0156&0.0040&0.0834&0.0101&0.9023&0.0106\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0540&0.0427&0.1045&0.0360&0.8479&0.0618\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0187&0.0045&0.0834&0.0101&0.8995&0.0108\\\hline \end{tabular} \begin{tabular}{|l|c|c|c|c|c|c|} \hline &\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline \textbf{BT-H}&0.0577&0.0213&0.1578&0.0295&0.7934&0.0262\\\hline \end{tabular}
2977094657/DidaAPI
1,997
frontend/docs/api/users.md
# 获取用户信息 获取当前登录用户的详细信息。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/user/profile` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 无需参数,获取当前登录用户的信息。 ## 响应格式 ### 成功响应 ```json { "etimestamp": null, "username": "string", "siteDomain": "dida365.com", "createdCampaign": "string", "createdDeviceInfo": null, "filledPassword": true, "accountDomain": "string", "extenalId": "string", "email": "string", "verifiedEmail": true, "fakedEmail": false, "phone": null, "name": "string", "givenName": null, "familyName": null, "link": null, "picture": "string", "gender": "string", "locale": "zh_CN", "userCode": "string", "verCode": null, "verKey": null, "externalId": "string", "phoneWithoutCountryCode": null, "displayName": "string" } ``` ## 响应字段说明 | 字段 | 类型 | 说明 | |------|------|------| | etimestamp | null | 时间戳 | | username | string | 用户名(通常是邮箱) | | siteDomain | string | 站点域名 | | createdCampaign | string | 创建活动 | | createdDeviceInfo | null | 创建设备信息 | | filledPassword | boolean | 是否已设置密码 | | accountDomain | string | 账户域名 | | extenalId | string | 外部ID | | email | string | 邮箱地址 | | verifiedEmail | boolean | 邮箱是否已验证 | | fakedEmail | boolean | 是否为虚假邮箱 | | phone | string/null | 手机号 | | name | string | 姓名 | | givenName | string/null | 名 | | familyName | string/null | 姓 | | link | string/null | 链接 | | picture | string | 头像URL | | gender | string | 性别(0-未知,1-男,2-女) | | locale | string | 语言设置 | | userCode | string | 用户代码 | | verCode | string/null | 验证码 | | verKey | string/null | 验证密钥 | | externalId | string | 外部标识 | | phoneWithoutCountryCode | string/null | 不含国家代码的手机号 | | displayName | string | 显示名称 | ## 使用说明 1. 确保已完成认证获取会话 2. 直接调用接口即可获取当前用户信息 3. 返回的数据包含用户的完整个人资料信息 4. 头像URL可直接用于显示用户头像 5. 用户代码(userCode)是用户的唯一标识 ## 应用场景 - **用户资料展示**: 在应用中显示用户的基本信息 - **头像显示**: 获取用户头像URL用于界面展示 - **用户身份验证**: 确认当前登录用户的身份 - **个人设置**: 获取用户的语言、性别等设置信息 - **账户管理**: 显示账户相关信息如邮箱验证状态
2977094657/DidaAPI
2,232
frontend/docs/api/projects.md
# 获取清单列表 获取用户的所有清单列表。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/projects` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 无需参数,使用当前认证会话。 ## 响应格式 ### 成功响应 ```json [ { "id": "string", "name": "string", "isOwner": true, "color": "string", "sortOrder": 0, "sortOption": { "groupBy": "string", "orderBy": "string" }, "sortType": "string", "userCount": 0, "etag": "string", "modifiedTime": "string", "inAll": true, "showType": null, "muted": true, "reminderType": null, "closed": null, "transferred": null, "groupId": null, "viewMode": "string", "notificationOptions": [ "string" ], "teamId": null, "permission": "string", "kind": "string", "timeline": { "range": null, "sortType": "string", "sortOption": { "groupBy": "string", "orderBy": "string" } }, "needAudit": true, "barcodeNeedAudit": true, "openToTeam": true, "teamMemberPermission": null, "source": 0 } ] ``` ## 响应字段说明 | 字段 | 类型 | 说明 | |------|------|------| | `id` | string | 清单唯一标识符 | | `name` | string | 清单名称 | | `isOwner` | boolean | 是否为拥有者 | | `color` | string | 清单颜色 | | `sortOrder` | number | 自定义排序值 | | `sortOption` | object | 排序选项配置 | | `sortType` | string | 排序类型 | | `userCount` | number | 用户数量 | | `etag` | string | 实体标签 | | `modifiedTime` | string | 修改时间 | | `inAll` | boolean | 是否在全部清单中显示 | | `showType` | any | 显示类型 | | `muted` | boolean | 是否静音 | | `reminderType` | any | 提醒类型 | | `closed` | any | 关闭状态 | | `transferred` | any | 转移状态 | | `groupId` | string/null | 分组ID | | `viewMode` | string | 视图模式 | | `notificationOptions` | array | 通知选项 | | `teamId` | string/null | 团队ID | | `permission` | string | 权限类型 | | `kind` | string | 清单类型 | | `timeline` | object | 时间线配置 | | `needAudit` | boolean | 是否需要审核 | | `barcodeNeedAudit` | boolean | 条码是否需要审核 | | `openToTeam` | boolean | 是否对团队开放 | | `teamMemberPermission` | any | 团队成员权限 | | `source` | number | 来源标识 |
2977094657/DidaAPI
1,715
frontend/docs/api/index.md
# 滴答清单API文档 本文档详细描述了滴答清单的原始API接口,帮助开发者了解如何直接调用滴答清单的服务。 ## 接口分类 ### 认证相关 - [微信登录流程](./auth/wechat-login-flow.md) - 完整的微信扫码登录流程 - [获取微信二维码](./auth/get-wechat-qrcode.md) - 获取微信登录二维码 - [轮询登录状态](./auth/poll-login-status.md) - 检查二维码扫码状态 - [验证微信登录](./auth/validate-wechat-login.md) - 验证微信登录并获取令牌 - [微信登录回调处理](./auth/wechat-callback.md) - 处理微信扫码后的回调,提取code参数 - [密码登录](./auth/password-login.md) - 使用用户名和密码进行登录 ### 任务管理 - [获取所有任务](./tasks/get-all-tasks.md) - 获取用户的所有任务列表 - [获取已完成任务](./tasks/get-completed-tasks.md) - 获取已完成/已放弃的任务列表,支持分页 - [获取垃圾桶任务](./tasks/get-trash-tasks.md) - 获取垃圾桶中的任务列表 - [获取任务统计](./tasks/get-tasks-summary.md) - 获取任务的统计信息 ### 清单管理 - [获取清单列表](./projects.md) - 获取用户的所有清单列表 ### 统计分析 - [获取用户排名统计](./statistics.md) - 获取用户在滴答清单中的排名和基本统计信息 - [获取通用统计信息](./statistics/general-statistics.md) - 获取概览、成就值、趋势等通用统计信息 - [获取任务统计信息](./statistics/task-statistics.md) - 获取指定日期范围内的任务统计信息 ### 番茄专注 - [获取番茄专注概览](./pomodoros.md) - 获取番茄专注的概览统计信息 ### 正计时专注 - [获取专注记录时间线](./pomodoros/focus-timeline.md) - 获取专注记录的时间线数据,支持分页 - [获取专注详情分布](./pomodoros/focus-distribution.md) - 获取指定日期范围内的专注时长分布统计 - [获取专注趋势热力图](./pomodoros/focus-heatmap.md) - 获取指定日期范围内的专注趋势热力图数据 - [获取专注时间按小时分布](./pomodoros/focus-hour-distribution.md) - 获取指定日期范围内按小时分布的专注时间统计 - [获取专注时间分布](./pomodoros/focus-time-distribution.md) - 获取指定日期范围内按时间段分布的专注数据 ### 习惯管理 - [获取所有习惯](./habits.md) - 获取当前用户的所有习惯列表 - [获取本周习惯打卡统计](./habits/week-current-statistics.md) - 获取本周的习惯打卡统计信息 - [导出习惯数据](./habits/export-habits.md) - 导出用户的习惯数据为Excel文件 ### 用户信息 - [获取用户信息](./users.md) - 获取当前登录用户的详细信息 ### 自定义接口 - [导出任务到Excel](./custom/export-tasks-excel.md) - 导出所有任务到Excel文件,包含全部任务、已完成任务、垃圾桶任务三个工作表 - [导出专注记录到Excel](./custom/export-focus-excel.md) - 导出所有专注记录到Excel文件,包含完整的专注时间线数据
294coder/Efficient-MIF
30,245
Pansharpening_Hyper_SR_Matlab_Test_Package/Demo1_Reduced_Resolution_WV3_MultiExm.m
%This is a demo to run fusion algorithms on Reduced Resolution % LJ Deng(UESTC) % 2022-06-02 clear; close all; %% =======load directors======== % Tools addpath([pwd,'/Tools']); % Select algorithms to run algorithms = {'GT'};%'PNN' location1 = [40 60 4 43]; % Location of zoom in location2 = []; %sensor = 'WV3'; %% =======read Multiple TestData_wv3.h5 (four 512x512 WV3 simulated data)======== file_test = '/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5'; disp(file_test) gt_multiExm_tmp = h5read(file_test,'/gt'); % WxHxCxN=1x2x3x4 gt_multiExm = permute(gt_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 ms_multiExm_tmp = h5read(file_test,'/ms'); % WxHxCxN=1x2x3x4 ms_multiExm = permute(ms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 lms_multiExm_tmp = h5read(file_test,'/lms'); % WxHxCxN=1x2x3x4 lms_multiExm = permute(lms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 pan_multiExm_tmp = h5read(file_test,'/pan'); % WxHxCxN=1x2x3x4 pan_multiExm = permute(pan_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 data_name = '3_EPS/WV3/multi/'; % director to save EPS figures %% ==========Read each Data==================== exm_num = size(ms_multiExm, 1); for i = 1 : exm_num % i = 1 or 2 ... %% read each data HRMS_tmp = gt_multiExm(i, :, :, :); % I_GT I_GT = squeeze(HRMS_tmp); LRMS_tmp = ms_multiExm(i, :, :, :); % I_MS_LR I_MS_LR = squeeze(LRMS_tmp); LMS_tmp = lms_multiExm(i, :, :, :); % I_MS I_MS = squeeze(LMS_tmp); PAN_tmp = pan_multiExm(i, :, :, :); % I_PAN I_PAN = squeeze(PAN_tmp); %% Initialization of the Matrix of Results NumIndexes = 5; MatrixResults = zeros(numel(algorithms),NumIndexes); alg = 0; %% load Indexes for WV3_RR sensor = 'GF2'; % disp(sensor) Qblocks_size = 32; bicubic = 0;% Interpolator flag_cut_bounds = 1;% Cut Final Image dim_cut = 30;% Cut Final Image thvalues = 0;% Threshold values out of dynamic range printEPS = 0;% Print Eps ratio = 4;% Resize Factor L = 11;% Radiometric Resolution %% show I_MS_LR, I_GT, PAN Imgs: % showImage8_zoomin(I_MS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % showPan(I_PAN,printEPS,2,flag_cut_bounds,dim_cut); % pause(2);print('-deps', strcat(data_name, num2str(i-1), '_pan', '.eps')) %% ======GT =================== if ismember('GT',algorithms) alg = alg + 1; [Q_avg_GT, SAM_GT, ERGAS_GT, SCC_GT, Q_GT] = indexes_evaluation(I_GT,I_MS,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_GT,Q_avg_GT,SAM_GT,ERGAS_GT,SCC_GT]; MatrixImage(:,:,:,alg) = I_GT; Q_avg_GT_multiexm(i) = Q_avg_GT; SAM_GT_multiexm(i) = SAM_GT; ERGAS_GT_multiexm(i) = ERGAS_GT; SCC_GT_multiexm(i) = SCC_GT; Q_GT_multiexm(i) = Q_GT; % showImage8_zoomin(I_GT,printEPS,2,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_gt', '.eps')) end %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%% CS-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% %% ====== 1) BT-H Method ====== if ismember('BT-H',algorithms) alg = alg + 1; cd BT-H t2=tic; I_BT_H = BroveyRegHazeMin(I_MS,I_PAN,ratio); time_BT_H = toc(t2); fprintf('Elaboration time BT-H: %.2f [sec]\n',time_BT_H); cd .. %%% Quality indexes computation [Q_avg_BT_H, SAM_BT_H, ERGAS_BT_H, SCC_BT_H, Q_BT_H] = indexes_evaluation(I_BT_H,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_BT_H,Q_avg_BT_H,SAM_BT_H,ERGAS_BT_H,SCC_BT_H]; MatrixImage(:,:,:,alg) = I_BT_H; Q_avg_BT_H_multiexm(i) = Q_avg_BT_H; SAM_BT_H_multiexm(i) = SAM_BT_H; ERGAS_BT_H_multiexm(i) = ERGAS_BT_H; SCC_BT_H_multiexm(i) = SCC_BT_H; Q_BT_H_multiexm(i) = Q_BT_H; % showImage8_zoomin(I_BT_H,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bth.eps')) end %% ====== 2) BDSD-PC Method ====== if ismember('BDSD-PC',algorithms) alg = alg + 1; cd BDSD t2=tic; I_BDSD_PC = BDSD_PC(I_MS,I_PAN,ratio,sensor); time_BDSD_PC = toc(t2); fprintf('Elaboration time BDSD-PC: %.2f [sec]\n',time_BDSD_PC); cd .. [Q_avg_BDSD_PC, SAM_BDSD_PC, ERGAS_BDSD_PC, SCC_BDSD_PC, Q_BDSD_PC] = indexes_evaluation(I_BDSD_PC,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_BDSD_PC,Q_avg_BDSD_PC,SAM_BDSD_PC,ERGAS_BDSD_PC,SCC_BDSD_PC]; MatrixImage(:,:,:,alg) = I_BDSD_PC; Q_avg_BDSD_PC_multiexm(i) = Q_avg_BDSD_PC; SAM_BDSD_PC_multiexm(i) = SAM_BDSD_PC; ERGAS_BDSD_PC_multiexm(i) = ERGAS_BDSD_PC; SCC_BDSD_PC_multiexm(i) = SCC_BDSD_PC; Q_BDSD_PC_multiexm(i) = Q_BDSD_PC; % showImage8_zoomin(I_BDSD_PC,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bdsd_pc.eps')) end %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%% MRA-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% %% ====== 1) MTF-GLP-HPM-R Method ====== if ismember('MTF-GLP-HPM-R',algorithms) alg = alg + 1; cd GLP t2=tic; I_MTF_GLP_HPM_R = MTF_GLP_HPM_R(I_MS,I_PAN,sensor,ratio); time_MTF_GLP_HPM_R = toc(t2); fprintf('Elaboration time MTF-GLP: %.2f [sec]\n',time_MTF_GLP_HPM_R); cd .. [Q_avg_MTF_GLP_HPM_R, SAM_MTF_GLP_HPM_R, ERGAS_MTF_GLP_HPM_R, SCC_MTF_GLP_HPM_R, Q_MTF_GLP_HPM_R] = indexes_evaluation(I_MTF_GLP_HPM_R,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_MTF_GLP_HPM_R,Q_avg_MTF_GLP_HPM_R,SAM_MTF_GLP_HPM_R,ERGAS_MTF_GLP_HPM_R,SCC_MTF_GLP_HPM_R]; MatrixImage(:,:,:,alg) = I_MTF_GLP_HPM_R; Q_avg_MTF_GLP_HPM_R_multiexm(i) = Q_avg_MTF_GLP_HPM_R; SAM_MTF_GLP_HPM_R_multiexm(i) = SAM_MTF_GLP_HPM_R; ERGAS_MTF_GLP_HPM_R_multiexm(i) = ERGAS_MTF_GLP_HPM_R; SCC_MTF_GLP_HPM_R_multiexm(i) = SCC_MTF_GLP_HPM_R; Q_MTF_GLP_HPM_R_multiexm(i) = Q_MTF_GLP_HPM_R; % showImage8_zoomin(I_MTF_GLP_HPM_R,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglp_hpm_r.eps')) end %% ====== 2) MTF-GLP-FS Method ====== if ismember('MTF-GLP-FS',algorithms) alg = alg + 1; cd GLP t2=tic; I_MTF_GLP_FS = MTF_GLP_FS(I_MS,I_PAN,sensor,ratio); time_MTF_GLP_FS = toc(t2); fprintf('Elaboration time MTF-GLP-FS: %.2f [sec]\n',time_MTF_GLP_FS); cd .. [Q_avg_MTF_GLP_FS, SAM_MTF_GLP_FS, ERGAS_MTF_GLP_FS, SCC_MTF_GLP_FS, Q_MTF_GLP_FS] = indexes_evaluation(I_MTF_GLP_FS,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_MTF_GLP_FS,Q_avg_MTF_GLP_FS,SAM_MTF_GLP_FS,ERGAS_MTF_GLP_FS,SCC_MTF_GLP_FS]; MatrixImage(:,:,:,alg) = I_MTF_GLP_FS; Q_avg_MTF_GLP_FS_multiexm(i) = Q_avg_MTF_GLP_FS; SAM_MTF_GLP_FS_multiexm(i) = SAM_MTF_GLP_FS; ERGAS_MTF_GLP_FS_multiexm(i) = ERGAS_MTF_GLP_FS; SCC_MTF_GLP_FS_multiexm(i) = SCC_MTF_GLP_FS; Q_MTF_GLP_FS_multiexm(i) = Q_MTF_GLP_FS; % showImage8_zoomin(I_MTF_GLP_FS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglpfs.eps')) end %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%% VO-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% %% ====== 1) TV Method ====== if ismember('TV',algorithms) alg = alg + 1; %%%%%%%%%%%%%%%%%%%%%%%%%% Parameters setting %%%%%%%%%%%%%%%%%%%%%%%%%%%%% switch sensor case 'IKONOS' w=[0.1091 0.2127 0.2928 0.3854]; c = 8; alpha=1.064; maxiter=10; lambda = 0.47106; case {'GeoEye1','WV4'} w=[0.1552, 0.3959, 0.2902, 0.1587]; c = 8; alpha=0.75; maxiter=50; lambda = 157.8954; case 'WV3' w=[0.0657 0.1012 0.1537 0.1473 0.1245 0.1545 0.1338 0.1192]; c = 8; alpha=0.75; maxiter=50; lambda = 1.0000e-03; end cd TV t2 = tic; I_TV = TV_pansharpen(I_MS_LR,I_PAN,alpha,lambda,c,maxiter,w); time_TV = toc(t2); fprintf('Elaboration time TV: %.2f [sec]\n',time_TV); cd .. [Q_avg_TV, SAM_TV, ERGAS_TV, SCC_TV, Q_TV] = indexes_evaluation(I_TV,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); MatrixResults(alg,:) = [Q_TV,Q_avg_TV,SAM_TV,ERGAS_TV,SCC_TV]; MatrixImage(:,:,:,alg) = I_TV; Q_avg_TV_multiexm(i) = Q_avg_TV; SAM_TV_multiexm(i) = SAM_TV; ERGAS_TV_multiexm(i) = ERGAS_TV; SCC_TV_multiexm(i) = SCC_TV; Q_TV_multiexm(i) = Q_TV; % showImage8_zoomin(I_TV,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_tv.eps')) end %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%% DL-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% %% ====== 1) PNN Method ====== % load(strcat('2_DL_Result/WV3_Reduced/PNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PNN % I_pnn = double(sr); % % if ismember('PNN',algorithms) % alg = alg + 1; % [Q_avg_pnn, SAM_pnn, ERGAS_pnn, SCC_pnn, Q_pnn] = indexes_evaluation(I_pnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_pnn,Q_avg_pnn,SAM_pnn,ERGAS_pnn,SCC_pnn]; % MatrixImage(:,:,:,alg) = I_pnn; % % Q_avg_pnn_multiexm(i) = Q_avg_pnn; % SAM_pnn_multiexm(i) = SAM_pnn; % ERGAS_pnn_multiexm(i) = ERGAS_pnn; % SCC_pnn_multiexm(i) = SCC_pnn; % Q_pnn_multiexm(i) = Q_pnn; % % showImage8_zoomin(I_pnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pnn.eps')) % end % %% ====== 2) PanNet Method ====== % % load(strcat('2_DL_Result/WV3_Reduced/PanNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PanNet % I_pannet = double(sr); % % if ismember('PanNet',algorithms) % alg = alg + 1; % [Q_avg_pannet, SAM_pannet, ERGAS_pannet, SCC_pannet, Q_pannet] = indexes_evaluation(I_pannet,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_pannet,Q_avg_pannet,SAM_pannet,ERGAS_pannet,SCC_pannet]; % MatrixImage(:,:,:,alg) = I_pannet; % % % Q_avg_pannet_multiexm(i) = Q_avg_pannet; % SAM_pannet_multiexm(i) = SAM_pannet; % ERGAS_pannet_multiexm(i) = ERGAS_pannet; % SCC_pannet_multiexm(i) = SCC_pannet; % Q_pannet_multiexm(i) = Q_pannet; % % showImage8_zoomin(I_pannet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pannet.eps')) % end %% ====== 3) DiCNN Method ====== % load(strcat('2_DL_Result/WV3_Reduced/DiCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for DiCNN % I_dicnn = double(sr); % % if ismember('DiCNN',algorithms) % alg = alg + 1; % [Q_avg_dicnn, SAM_dicnn, ERGAS_dicnn, SCC_dicnn, Q_dicnn] = indexes_evaluation(I_dicnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_dicnn,Q_avg_dicnn,SAM_dicnn,ERGAS_dicnn,SCC_dicnn]; % MatrixImage(:,:,:,alg) = I_dicnn; % % Q_avg_dicnn_multiexm(i) = Q_avg_dicnn; % SAM_dicnn_multiexm(i) = SAM_dicnn; % ERGAS_dicnn_multiexm(i) = ERGAS_dicnn; % SCC_dicnn_multiexm(i) = SCC_dicnn; % Q_dicnn_multiexm(i) = Q_dicnn; % % showImage8_zoomin(I_dicnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_dicnn.eps')) % end % % %% ====== 4) MSDCNN Method ====== % load(strcat('2_DL_Result/WV3_Reduced/MSDCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for MSDCNN % I_msdcnn = double(sr); % % if ismember('MSDCNN',algorithms) % alg = alg + 1; % [Q_avg_msdcnn, SAM_msdcnn, ERGAS_msdcnn, SCC_msdcnn, Q_msdcnn] = indexes_evaluation(I_msdcnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_msdcnn,Q_avg_msdcnn,SAM_msdcnn,ERGAS_msdcnn,SCC_msdcnn]; % MatrixImage(:,:,:,alg) = I_msdcnn; % % Q_avg_msdcnn_multiexm(i) = Q_avg_msdcnn; % SAM_msdcnn_multiexm(i) = SAM_msdcnn; % ERGAS_msdcnn_multiexm(i) = ERGAS_msdcnn; % SCC_msdcnn_multiexm(i) = SCC_msdcnn; % Q_msdcnn_multiexm(i) = Q_msdcnn; % % showImage8_zoomin(I_msdcnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_msdcnn.eps')) % end % %% ====== 5) BDPN Method ====== % load(strcat('2_DL_Result/WV3_Reduced/BDPN/results/output_mulExm_' , num2str(i-1), '.mat')) % load i-th image for BDPN % I_bdpn = double(sr); % % if ismember('BDPN',algorithms) % alg = alg + 1; % [Q_avg_bdpn, SAM_bdpn, ERGAS_bdpn, SCC_bdpn, Q_bdpn] = indexes_evaluation(I_bdpn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_bdpn,Q_avg_bdpn,SAM_bdpn,ERGAS_bdpn,SCC_bdpn]; % MatrixImage(:,:,:,alg) = I_bdpn; % % Q_avg_bdpn_multiexm(i) = Q_avg_bdpn; % SAM_bdpn_multiexm(i) = SAM_bdpn; % ERGAS_bdpn_multiexm(i) = ERGAS_bdpn; % SCC_bdpn_multiexm(i) = SCC_bdpn; % Q_bdpn_multiexm(i) = Q_bdpn; % % showImage8_zoomin(I_bdpn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_bdpn.eps')) % end % % %% ====== 6) FusionNet Method ====== % load(strcat('2_DL_Result/WV3_Reduced/FusionNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for FusionNet % I_fusionnet = double(sr); % % if ismember('FusionNet',algorithms) % alg = alg + 1; % [Q_avg_fusionnet, SAM_fusionnet, ERGAS_fusionnet, SCC_fusionnet, Q_fusionnet] = indexes_evaluation(I_fusionnet,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_fusionnet,Q_avg_fusionnet,SAM_fusionnet,ERGAS_fusionnet,SCC_fusionnet]; % MatrixImage(:,:,:,alg) = I_fusionnet; % % Q_avg_fusionnet_multiexm(i) = Q_avg_fusionnet; % SAM_fusionnet_multiexm(i) = SAM_fusionnet; % ERGAS_fusionnet_multiexm(i) = ERGAS_fusionnet; % SCC_fusionnet_multiexm(i) = SCC_fusionnet; % Q_fusionnet_multiexm(i) = Q_fusionnet; % % showImage8_zoomin(I_fusionnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_fusionnet.eps')) % end %% ====== 7) LAGConv Method ====== % load(strcat('2_DL_Result/WV3_Reduced/LAGConv/results/output_mulExm_',num2str(i-1), '.mat')) % load i-th image for LAGConv % I_lagnet = double(sr); % % if ismember('LAGConv',algorithms) % alg = alg + 1; % [Q_avg_lagnet, SAM_lagnet, ERGAS_lagnet, SCC_lagnet, Q_lagnet] = indexes_evaluation(I_lagnet,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues); % MatrixResults(alg,:) = [Q_lagnet,Q_avg_lagnet,SAM_lagnet,ERGAS_lagnet,SCC_lagnet]; % MatrixImage(:,:,:,alg) = I_lagnet; % % Q_avg_lagnet_multiexm(i) = Q_avg_lagnet; % SAM_lagnet_multiexm(i) = SAM_lagnet; % ERGAS_lagnet_multiexm(i) = ERGAS_lagnet; % SCC_lagnet_multiexm(i) = SCC_lagnet; % Q_lagnet_multiexm(i) = Q_lagnet; % % showImage8_zoomin(I_lagnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_lagnet.eps')) % end end %% Print in LATEX %matrix2latex(MatrixResults(:,[1,3,4]),'RR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'Q2n'},{'SAM'},{'ERGAS'}],'alignment','c','format', '%.4f'); %% View All if size(I_GT,3) == 4 vect_index_RGB = [3,2,1]; else vect_index_RGB = [5,3,2]; end titleImages = algorithms; figure, showImagesAll(MatrixImage,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0); %% ======Display the final average performance ======= % GT: average Q_avg avg_Q_GT_multiexm = mean(Q_GT_multiexm); std_Q_GT_multiexm = std(Q_GT_multiexm); avg_Q_avg_GT_multiexm = mean(Q_avg_GT_multiexm); std_Q_avg_GT_multiexm = std(Q_avg_GT_multiexm); avg_SAM_GT_multiexm = mean(SAM_GT_multiexm); std_SAM_GT_multiexm = std(SAM_GT_multiexm); avg_ERGAS_GT_multiexm = mean(ERGAS_GT_multiexm); std_ERGAS_GT_multiexm = std(ERGAS_GT_multiexm); avg_SCC_GT_multiexm = mean(SCC_GT_multiexm); std_SCC_GT_multiexm = std(SCC_GT_multiexm); Avg_MatrixResults(1,:) = [avg_Q_GT_multiexm, std_Q_GT_multiexm, avg_Q_avg_GT_multiexm, std_Q_avg_GT_multiexm, ... avg_SAM_GT_multiexm, std_SAM_GT_multiexm, avg_ERGAS_GT_multiexm, std_ERGAS_GT_multiexm,... avg_SCC_GT_multiexm, std_SCC_GT_multiexm]; % BT_H: average Q_avg % avg_Q_BT_H_multiexm = mean(Q_BT_H_multiexm); % std_Q_BT_H_multiexm = std(Q_BT_H_multiexm); % avg_Q_avg_BT_H_multiexm = mean(Q_avg_BT_H_multiexm); % std_Q_avg_BT_H_multiexm = std(Q_avg_BT_H_multiexm); % avg_SAM_BT_H_multiexm = mean(SAM_BT_H_multiexm); % std_SAM_BT_H_multiexm = std(SAM_BT_H_multiexm); % avg_ERGAS_BT_H_multiexm = mean(ERGAS_BT_H_multiexm); % std_ERGAS_BT_H_multiexm = std(ERGAS_BT_H_multiexm); % avg_SCC_BT_H_multiexm = mean(SCC_BT_H_multiexm); % std_SCC_BT_H_multiexm = std(SCC_BT_H_multiexm); % Avg_MatrixResults(2,:) = [avg_Q_BT_H_multiexm, std_Q_BT_H_multiexm, avg_Q_avg_BT_H_multiexm, std_Q_avg_BT_H_multiexm, ... % avg_SAM_BT_H_multiexm, std_SAM_BT_H_multiexm, avg_ERGAS_BT_H_multiexm, std_ERGAS_BT_H_multiexm,... % avg_SCC_BT_H_multiexm, std_SCC_BT_H_multiexm]; % % BDSD: average Q_avg % avg_Q_BDSD_PC_multiexm = mean(Q_BDSD_PC_multiexm); % std_Q_BDSD_PC_multiexm = std(Q_BDSD_PC_multiexm); % avg_Q_avg_BDSD_PC_multiexm = mean(Q_avg_BDSD_PC_multiexm); % std_Q_avg_BDSD_PC_multiexm = std(Q_avg_BDSD_PC_multiexm); % avg_SAM_BDSD_PC_multiexm = mean(SAM_BDSD_PC_multiexm); % std_SAM_BDSD_PC_multiexm = std(SAM_BDSD_PC_multiexm); % avg_ERGAS_BDSD_PC_multiexm = mean(ERGAS_BDSD_PC_multiexm); % std_ERGAS_BDSD_PC_multiexm = std(ERGAS_BDSD_PC_multiexm); % avg_SCC_BDSD_PC_multiexm = mean(SCC_BDSD_PC_multiexm); % std_SCC_BDSD_PC_multiexm = std(SCC_BDSD_PC_multiexm); % Avg_MatrixResults(3,:) = [avg_Q_BDSD_PC_multiexm, std_Q_BDSD_PC_multiexm, avg_Q_avg_BDSD_PC_multiexm, std_Q_avg_BDSD_PC_multiexm, ... % avg_SAM_BDSD_PC_multiexm, std_SAM_BDSD_PC_multiexm, avg_ERGAS_BDSD_PC_multiexm, std_ERGAS_BDSD_PC_multiexm,... % avg_SCC_BDSD_PC_multiexm, std_SCC_BDSD_PC_multiexm]; % % MTF_GLP: average Q_avg % avg_Q_MTF_GLP_HPM_R_multiexm = mean(Q_MTF_GLP_HPM_R_multiexm); % std_Q_MTF_GLP_HPM_R_multiexm = std(Q_MTF_GLP_HPM_R_multiexm); % avg_Q_avg_MTF_GLP_HPM_R_multiexm = mean(Q_avg_MTF_GLP_HPM_R_multiexm); % std_Q_avg_MTF_GLP_HPM_R_multiexm = std(Q_avg_MTF_GLP_HPM_R_multiexm); % avg_SAM_MTF_GLP_HPM_R_multiexm = mean(SAM_MTF_GLP_HPM_R_multiexm); % std_SAM_MTF_GLP_HPM_R_multiexm = std(SAM_MTF_GLP_HPM_R_multiexm); % avg_ERGAS_MTF_GLP_HPM_R_multiexm = mean(ERGAS_MTF_GLP_HPM_R_multiexm); % std_ERGAS_MTF_GLP_HPM_R_multiexm = std(ERGAS_MTF_GLP_HPM_R_multiexm); % avg_SCC_MTF_GLP_HPM_R_multiexm = mean(SCC_MTF_GLP_HPM_R_multiexm); % std_SCC_MTF_GLP_HPM_R_multiexm = std(SCC_MTF_GLP_HPM_R_multiexm); % Avg_MatrixResults(4,:) = [avg_Q_MTF_GLP_HPM_R_multiexm, std_Q_MTF_GLP_HPM_R_multiexm, avg_Q_avg_MTF_GLP_HPM_R_multiexm, std_Q_avg_MTF_GLP_HPM_R_multiexm, ... % avg_SAM_MTF_GLP_HPM_R_multiexm, std_SAM_MTF_GLP_HPM_R_multiexm, avg_ERGAS_MTF_GLP_HPM_R_multiexm, std_ERGAS_MTF_GLP_HPM_R_multiexm,... % avg_SCC_MTF_GLP_HPM_R_multiexm, std_SCC_MTF_GLP_HPM_R_multiexm]; % % MTF_GLP_FS: average Q_avg % avg_Q_MTF_GLP_FS_multiexm = mean(Q_MTF_GLP_FS_multiexm); % std_Q_MTF_GLP_FS_multiexm = std(Q_MTF_GLP_FS_multiexm); % avg_Q_avg_MTF_GLP_FS_multiexm = mean(Q_avg_MTF_GLP_FS_multiexm); % std_Q_avg_MTF_GLP_FS_multiexm = std(Q_avg_MTF_GLP_FS_multiexm); % avg_SAM_MTF_GLP_FS_multiexm = mean(SAM_MTF_GLP_FS_multiexm); % std_SAM_MTF_GLP_FS_multiexm = std(SAM_MTF_GLP_FS_multiexm); % avg_ERGAS_MTF_GLP_FS_multiexm = mean(ERGAS_MTF_GLP_FS_multiexm); % std_ERGAS_MTF_GLP_FS_multiexm = std(ERGAS_MTF_GLP_FS_multiexm); % avg_SCC_MTF_GLP_FS_multiexm = mean(SCC_MTF_GLP_FS_multiexm); % std_SCC_MTF_GLP_FS_multiexm = std(SCC_MTF_GLP_FS_multiexm); % Avg_MatrixResults(5,:) = [avg_Q_MTF_GLP_FS_multiexm, std_Q_MTF_GLP_FS_multiexm, avg_Q_avg_MTF_GLP_FS_multiexm, std_Q_avg_MTF_GLP_FS_multiexm, ... % avg_SAM_MTF_GLP_FS_multiexm, std_SAM_MTF_GLP_FS_multiexm, avg_ERGAS_MTF_GLP_FS_multiexm, std_ERGAS_MTF_GLP_FS_multiexm,... % avg_SCC_MTF_GLP_FS_multiexm, std_SCC_MTF_GLP_FS_multiexm]; % % TV: average Q_avg % avg_Q_TV_multiexm = mean(Q_TV_multiexm); % std_Q_TV_multiexm = std(Q_TV_multiexm); % avg_Q_avg_TV_multiexm = mean(Q_avg_TV_multiexm); % std_Q_avg_TV_multiexm = std(Q_avg_TV_multiexm); % avg_SAM_TV_multiexm = mean(SAM_TV_multiexm); % std_SAM_TV_multiexm = std(SAM_TV_multiexm); % avg_ERGAS_TV_multiexm = mean(ERGAS_TV_multiexm); % std_ERGAS_TV_multiexm = std(ERGAS_TV_multiexm); % avg_SCC_TV_multiexm = mean(SCC_TV_multiexm); % std_SCC_TV_multiexm = std(SCC_TV_multiexm); % Avg_MatrixResults(6,:) = [avg_Q_TV_multiexm, std_Q_TV_multiexm, avg_Q_avg_TV_multiexm, std_Q_avg_TV_multiexm, ... % avg_SAM_TV_multiexm, std_SAM_TV_multiexm, avg_ERGAS_TV_multiexm, std_ERGAS_TV_multiexm,... % avg_SCC_TV_multiexm, std_SCC_TV_multiexm]; % pnn: average Q_avg %avg_Q_pnn_multiexm = mean(Q_pnn_multiexm); %std_Q_pnn_multiexm = std(Q_pnn_multiexm); % %avg_Q_avg_pnn_multiexm = mean(Q_avg_pnn_multiexm); %std_Q_avg_pnn_multiexm = std(Q_avg_pnn_multiexm); % %avg_SAM_pnn_multiexm = mean(SAM_pnn_multiexm); %std_SAM_pnn_multiexm = std(SAM_pnn_multiexm); % %avg_ERGAS_pnn_multiexm = mean(ERGAS_pnn_multiexm); %std_ERGAS_pnn_multiexm = std(ERGAS_pnn_multiexm); % %avg_SCC_pnn_multiexm = mean(SCC_pnn_multiexm); %std_SCC_pnn_multiexm = std(SCC_pnn_multiexm); % %Avg_MatrixResults(7,:) = [avg_Q_pnn_multiexm, std_Q_pnn_multiexm, avg_Q_avg_pnn_multiexm, std_Q_avg_pnn_multiexm, ... % avg_SAM_pnn_multiexm, std_SAM_pnn_multiexm, avg_ERGAS_pnn_multiexm, std_ERGAS_pnn_multiexm,... % avg_SCC_pnn_multiexm, std_SCC_pnn_multiexm]; % pannet: average Q_avg %avg_Q_pannet_multiexm = mean(Q_pannet_multiexm); %std_Q_pannet_multiexm = std(Q_pannet_multiexm); % %avg_Q_avg_pannet_multiexm = mean(Q_avg_pannet_multiexm); %std_Q_avg_pannet_multiexm = std(Q_avg_pannet_multiexm); % %avg_SAM_pannet_multiexm = mean(SAM_pannet_multiexm); %std_SAM_pannet_multiexm = std(SAM_pannet_multiexm); % %avg_ERGAS_pannet_multiexm = mean(ERGAS_pannet_multiexm); %std_ERGAS_pannet_multiexm = std(ERGAS_pannet_multiexm); % %avg_SCC_pannet_multiexm = mean(SCC_pannet_multiexm); %std_SCC_pannet_multiexm = std(SCC_pannet_multiexm); % %Avg_MatrixResults(8,:) = [avg_Q_pannet_multiexm, std_Q_pannet_multiexm, avg_Q_avg_pannet_multiexm, std_Q_avg_pannet_multiexm, ... % avg_SAM_pannet_multiexm, std_SAM_pannet_multiexm, avg_ERGAS_pannet_multiexm, std_ERGAS_pannet_multiexm,... % avg_SCC_pannet_multiexm, std_SCC_pannet_multiexm]; % %% dicnn: average Q_avg %avg_Q_dicnn_multiexm = mean(Q_dicnn_multiexm); %std_Q_dicnn_multiexm = std(Q_dicnn_multiexm); % %avg_Q_avg_dicnn_multiexm = mean(Q_avg_dicnn_multiexm); %std_Q_avg_dicnn_multiexm = std(Q_avg_dicnn_multiexm); % %avg_SAM_dicnn_multiexm = mean(SAM_dicnn_multiexm); %std_SAM_dicnn_multiexm = std(SAM_dicnn_multiexm); % %avg_ERGAS_dicnn_multiexm = mean(ERGAS_dicnn_multiexm); %std_ERGAS_dicnn_multiexm = std(ERGAS_dicnn_multiexm); % %avg_SCC_dicnn_multiexm = mean(SCC_dicnn_multiexm); %std_SCC_dicnn_multiexm = std(SCC_dicnn_multiexm); % %Avg_MatrixResults(9,:) = [avg_Q_dicnn_multiexm, std_Q_dicnn_multiexm, avg_Q_avg_dicnn_multiexm, std_Q_avg_dicnn_multiexm, ... % avg_SAM_dicnn_multiexm, std_SAM_dicnn_multiexm, avg_ERGAS_dicnn_multiexm, std_ERGAS_dicnn_multiexm,... % avg_SCC_dicnn_multiexm, std_SCC_dicnn_multiexm]; % % % %% %% msdcnn: average Q_avg %avg_Q_msdcnn_multiexm = mean(Q_msdcnn_multiexm); %std_Q_msdcnn_multiexm = std(Q_msdcnn_multiexm); % %avg_Q_avg_msdcnn_multiexm = mean(Q_avg_msdcnn_multiexm); %std_Q_avg_msdcnn_multiexm = std(Q_avg_msdcnn_multiexm); % %avg_SAM_msdcnn_multiexm = mean(SAM_msdcnn_multiexm); %std_SAM_msdcnn_multiexm = std(SAM_msdcnn_multiexm); % %avg_ERGAS_msdcnn_multiexm = mean(ERGAS_msdcnn_multiexm); %std_ERGAS_msdcnn_multiexm = std(ERGAS_msdcnn_multiexm); % %avg_SCC_msdcnn_multiexm = mean(SCC_msdcnn_multiexm); %std_SCC_msdcnn_multiexm = std(SCC_msdcnn_multiexm); % %Avg_MatrixResults(10,:) = [avg_Q_msdcnn_multiexm, std_Q_msdcnn_multiexm, avg_Q_avg_msdcnn_multiexm, std_Q_avg_msdcnn_multiexm, ... % avg_SAM_msdcnn_multiexm, std_SAM_msdcnn_multiexm, avg_ERGAS_msdcnn_multiexm, std_ERGAS_msdcnn_multiexm,... % avg_SCC_msdcnn_multiexm, std_SCC_msdcnn_multiexm]; % % %% bdpn: average Q_avg %avg_Q_bdpn_multiexm = mean(Q_bdpn_multiexm); %std_Q_bdpn_multiexm = std(Q_bdpn_multiexm); % %avg_Q_avg_bdpn_multiexm = mean(Q_avg_bdpn_multiexm); %std_Q_avg_bdpn_multiexm = std(Q_avg_bdpn_multiexm); % %avg_SAM_bdpn_multiexm = mean(SAM_bdpn_multiexm); %std_SAM_bdpn_multiexm = std(SAM_bdpn_multiexm); % %avg_ERGAS_bdpn_multiexm = mean(ERGAS_bdpn_multiexm); %std_ERGAS_bdpn_multiexm = std(ERGAS_bdpn_multiexm); % %avg_SCC_bdpn_multiexm = mean(SCC_bdpn_multiexm); %std_SCC_bdpn_multiexm = std(SCC_bdpn_multiexm); % %Avg_MatrixResults(11,:) = [avg_Q_bdpn_multiexm, std_Q_bdpn_multiexm, avg_Q_avg_bdpn_multiexm, std_Q_avg_bdpn_multiexm, ... % avg_SAM_bdpn_multiexm, std_SAM_bdpn_multiexm, avg_ERGAS_bdpn_multiexm, std_ERGAS_bdpn_multiexm,... % avg_SCC_bdpn_multiexm, std_SCC_bdpn_multiexm]; % % % %% fusionnet: average Q_avg %avg_Q_fusionnet_multiexm = mean(Q_fusionnet_multiexm); %std_Q_fusionnet_multiexm = std(Q_fusionnet_multiexm); % %avg_Q_avg_fusionnet_multiexm = mean(Q_avg_fusionnet_multiexm); %std_Q_avg_fusionnet_multiexm = std(Q_avg_fusionnet_multiexm); % %avg_SAM_fusionnet_multiexm = mean(SAM_fusionnet_multiexm); %std_SAM_fusionnet_multiexm = std(SAM_fusionnet_multiexm); % %avg_ERGAS_fusionnet_multiexm = mean(ERGAS_fusionnet_multiexm); %std_ERGAS_fusionnet_multiexm = std(ERGAS_fusionnet_multiexm); % %avg_SCC_fusionnet_multiexm = mean(SCC_fusionnet_multiexm); %std_SCC_fusionnet_multiexm = std(SCC_fusionnet_multiexm); % %Avg_MatrixResults(12,:) = [avg_Q_fusionnet_multiexm, std_Q_fusionnet_multiexm, avg_Q_avg_fusionnet_multiexm, std_Q_avg_fusionnet_multiexm, ... % avg_SAM_fusionnet_multiexm, std_SAM_fusionnet_multiexm, avg_ERGAS_fusionnet_multiexm, std_ERGAS_fusionnet_multiexm,... % avg_SCC_fusionnet_multiexm, std_SCC_fusionnet_multiexm]; %% lagnet: average Q_avg %avg_Q_lagnet_multiexm = mean(Q_lagnet_multiexm); %std_Q_lagnet_multiexm = std(Q_lagnet_multiexm); % %avg_Q_avg_lagnet_multiexm = mean(Q_avg_lagnet_multiexm); %std_Q_avg_lagnet_multiexm = std(Q_avg_lagnet_multiexm); % %avg_SAM_lagnet_multiexm = mean(SAM_lagnet_multiexm); %std_SAM_lagnet_multiexm = std(SAM_lagnet_multiexm); % %avg_ERGAS_lagnet_multiexm = mean(ERGAS_lagnet_multiexm); %std_ERGAS_lagnet_multiexm = std(ERGAS_lagnet_multiexm); % %avg_SCC_lagnet_multiexm = mean(SCC_lagnet_multiexm); %std_SCC_lagnet_multiexm = std(SCC_lagnet_multiexm); % %Avg_MatrixResults(13,:) = [avg_Q_lagnet_multiexm, std_Q_lagnet_multiexm, avg_Q_avg_lagnet_multiexm, std_Q_avg_lagnet_multiexm, ... % avg_SAM_lagnet_multiexm, std_SAM_lagnet_multiexm, avg_ERGAS_lagnet_multiexm, std_ERGAS_lagnet_multiexm,... % avg_SCC_lagnet_multiexm, std_SCC_lagnet_multiexm]; %matrix2latex(Avg_MatrixResults(:,[1,2, 5,6, 7,8 ]),'Avg_RR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'Q2n'}, {'Q2n-std'}, {'SAM'}, {'SAM-std'}, {'ERGAS'}, {'ERGAS-std'}],'alignment','c','format', '%.4f'); fprintf('\n') disp('#######################################################') %disp(['Display the performance for:', num2str(1:i)]) disp('#######################################################') disp(' |====Q====|===Q_avg===|=====SAM=====|======ERGAS=======|=======SCC=======') Avg_MatrixResults
294coder/Efficient-MIF
1,176
Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_unref_batched_images.m
function res = analysis_unref_batched_images(path, ratio, sensor) if strcmp(sensor, 'QB') || strcmp(sensor, 'GF2') sensor = 'IKONOS'; end disp(sensor) data=load(path); sr = data.sr; ms = data.ms; lms = data.lms; pan = data.pan; addpath('./Quality_Indices/') addpath('./Tools') sz = size(sr); bs = sz(1); d_lambdas = []; qnr_indices = []; d_ses = []; blockSize = 32; res = {}; h = size(pan, 3); w = size(pan, 4); for i = (1: bs) sr1 = permute(squeeze(sr(i, :, :, :)), [2, 3, 1]); ms1 = permute(squeeze(ms(i, :, :, :)), [2, 3, 1]); lms1 = permute(squeeze(lms(i, :, :, :)), [2, 3, 1]); pan1 = permute(reshape(squeeze(pan(i, :, :, :)), 1, h, w), [2, 3, 1]); [d_lambdas(i), d_ses(i), qnr_indices(i)] = indexes_evaluation_FS(sr1, ms1, pan1, 11, 0, lms1, sensor, ratio, 0); fprintf('sample %d - d_lambda: %f, qnr_index: %f, d_s: %f \n', i, d_lambdas(i), qnr_indices(i), d_ses(i)) end res.d_lambda = [mean(d_lambdas), std(d_lambdas)]; res.qnr_index = [mean(qnr_indices), std(qnr_indices)]; res.d_s = [mean(d_ses), std(d_ses)]; end
294coder/Efficient-MIF
1,748
Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_unref_unbatched_images.m
function res = analysis_unref_unbatched_images(dir_path, ratio, sensor) full_data_path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_wv3_OrigScale_multiExm1.h5"; sensor = lower(sensor); % if sensor == "wv3" || sensor == "wv2" || sensor == "qb" % const = 2047; % elseif sensor == "gf2" || sensor == "gf" % const = 1023; % else % error(strcat(sensor, ' is not supported!')) ms_s = h5read(full_data_path, '/ms'); ms_s = permute(ms_s, [4,2,1,3]); lms_s = h5read(full_data_path, '/lms'); lms_s = permute(lms_s, [4,2,1,3]); pan_s = h5read(full_data_path, '/pan'); pan_s = permute(pan_s, [4,2,1,3]); bs = size(lms_s, 1); addpath('./Tools') addpath('./Quality_Indices/') d_lambdas = []; qnr_indices = []; d_ses = []; res = {}; h = size(pan_s, 3); w = size(pan_s, 4); for i = (0:bs-1) p = strcat(dir_path, "/", "output_mulExm_", string(i), ".mat"); sr2 = load(p); sr = sr2.sr; i = i+1; ms = squeeze(ms_s(i, :, :, :)); lms = squeeze(lms_s(i, :, :, :)); pan = squeeze(pan_s(i, :, :, :)); [d_lambdas(i), d_ses(i), qnr_indices(i)] = indexes_evaluation_FS(sr, ms, pan, 11, 0, lms, sensor, ratio, 0); fprintf('sample %d - d_lambda: %f, qnr_index: %f, d_s: %f \n', i, d_lambdas(i), qnr_indices(i), d_ses(i)) res.d_lambda = [mean(d_lambdas), std(d_lambdas)]; res.qnr_index = [mean(qnr_indices), std(qnr_indices)]; res.d_s = [mean(d_ses), std(d_ses)]; end res.d_lambda = [mean(d_lambdas), std(d_lambdas)]; res.qnr_index = [mean(qnr_indices), std(qnr_indices)]; res.d_s = [mean(d_ses), std(d_ses)]; end
2977094657/DidaAPI
1,142
frontend/docs/api/habits.md
# 获取所有习惯 获取当前用户的所有习惯列表。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/habits` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 无需参数,使用当前认证会话。 ## 响应格式 ### 成功响应 ```json [ { "id": "string", "name": "string", "iconRes": "string", "color": "string", "status": 1, "encouragement": "string", "totalCheckIns": 0, "createdTime": "string", "modifiedTime": "string", "type": "string", "goal": 0.0 }, { "id": "string", "name": "string", "iconRes": "string", "color": "string", "status": 0, "encouragement": "string", "totalCheckIns": 0, "createdTime": "string", "modifiedTime": "string", "type": "string", "goal": 0.0 } ] ``` ## 响应字段说明 | 字段 | 类型 | 说明 | |------|------|------| | id | string | 习惯ID | | name | string | 习惯名称 | | iconRes | string | 图标资源名称 | | color | string | 习惯颜色(十六进制) | | status | number | 状态(0=未完成,1=已完成) | | encouragement | string | 激励语句 | | totalCheckIns | number | 总打卡次数 | | createdTime | string | 创建时间 | | modifiedTime | string | 修改时间 | | type | string | 习惯类型(daily/weekly等) | | goal | number | 目标值 |
2977094657/DidaAPI
2,293
frontend/docs/api/url-management.md
# URL和外部链接管理 本项目采用统一的URL管理策略,所有外部链接和API端点都在 `core/urls.py` 文件中集中管理。 ## 管理原则 ### 为什么需要统一管理 1. **避免硬编码** - 防止URL散布在各个文件中 2. **便于维护** - 统一修改和更新链接 3. **提高可读性** - 清晰的分类和注释 4. **便于测试** - 集中的URL便于健康检查 ### 分类管理 所有URL按功能分为以下几类: #### 微信开放平台相关 - `WECHAT_URLS` - 微信API的基础URL - `WECHAT_CONFIG` - 微信应用配置参数 #### 滴答清单API相关 - `DIDA_API_BASE` - 滴答清单API基础配置 - `DIDA_AUTH_APIS` - 认证相关API端点 - `DIDA_TASK_APIS` - 任务管理API端点 - `DIDA_PROJECT_APIS` - 项目管理API端点 #### 外部文档链接 - `OFFICIAL_DOCS` - 官方文档链接 - `TECH_REFERENCES` - 技术参考文档 ## 使用方法 ### 导入URL模块 ```python from core import urls ``` ### 使用预定义的URL ```python # 获取微信二维码基础URL qr_base_url = urls.WECHAT_URLS["qr_base_url"] # 获取滴答清单API基础URL api_base = urls.DIDA_API_BASE["base_url"] ``` ### 使用URL构建函数 ```python # 构建微信登录二维码URL qr_url = urls.build_wechat_qr_url(state="Lw==") # 构建滴答清单API完整URL api_url = urls.build_dida_api_url("/batch/check/0") # 构建微信登录验证URL validate_url = urls.build_wechat_validate_url(code="xxx", state="Lw==") ``` ## URL分类详情 ### 微信相关URL ```python WECHAT_URLS = { "qr_base_url": "https://open.weixin.qq.com/connect/qrconnect", "qr_image_base_url": "https://open.weixin.qq.com/connect/qrcode", "poll_login_url": "https://long.open.weixin.qq.com/connect/l/qrconnect", "redirect_uri": "https://dida365.com/sign/wechat" } ``` ### 滴答清单API端点 ```python DIDA_TASK_APIS = { "get_all_tasks": "/batch/check/0", "task_crud": "/task", "task_search": "/task/search" } ``` ### 官方文档链接 ```python OFFICIAL_DOCS = { "wechat_login_guide": "https://developers.weixin.qq.com/doc/oplatform/Website_App/WeChat_Login/Wechat_Login", "dida_official": "https://dida365.com" } ``` ## 辅助函数 ### URL构建函数 - `build_wechat_qr_url(state)` - 构建微信登录二维码URL - `build_wechat_poll_url(uuid, timestamp)` - 构建轮询URL - `build_dida_api_url(endpoint)` - 构建滴答清单API URL - `build_wechat_validate_url(code, state)` - 构建验证URL ### 管理函数 - `get_all_external_urls()` - 获取所有外部URL - `get_api_endpoints()` - 获取所有API端点 ## 添加新URL的步骤 ### 1. 确定分类 根据URL的用途选择合适的分类字典。 ### 2. 添加URL ```python # 在对应的字典中添加新URL DIDA_TASK_APIS = { # 现有URL... "new_endpoint": "/new/api/endpoint" # 新增 } ``` ### 3. 添加注释 为新URL添加清晰的注释说明其用途。 ### 4. 更新文档 在相关的API文档中引用新的URL配置。 ## 相关文件 - `core/urls.py` - URL管理主文件 - `core/config.py` - 非URL配置管理 - `config.toml` - 配置文件(已移除URL配置)
294coder/Efficient-MIF
23,781
Pansharpening_Hyper_SR_Matlab_Test_Package/Demo2_Full_Resolution_WV3_multi.m
% LJ Deng(UESTC) % 2020-06-02 clear; close all; %% =======load directors======== % Tools addpath([pwd,'/Tools']); % L, locatio % % Select algorithms to run algorithms = {'BT-H'};%'PNN' data_name = '3_EPS/WV3/wv3_os_'; % director to save EPS figures %% ==========Read each Data==================== %% read each data file_test = '/Data2/ZiHanCao/datasets/pansharpening/qb/full_examples/test_qb_OrigScale_multiExm1.h5'; disp(file_test) %load(file_test) % get I_MS_LR, I_MS, I_PAN and sensors' info. ms_multiExm_tmp = h5read(file_test,'/ms'); % WxHxCxN=1x2x3x4 ms_multiExm = permute(ms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 lms_multiExm_tmp = h5read(file_test,'/lms'); % WxHxCxN=1x2x3x4 lms_multiExm = permute(lms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 pan_multiExm_tmp = h5read(file_test,'/pan'); % WxHxCxN=1x2x3x4 pan_multiExm = permute(pan_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3 %% ==========Read each Data==================== exm_num = size(ms_multiExm, 1); for i = (1: exm_num) disp(i) %% read each data LRMS_tmp = ms_multiExm(i, :, :, :); % I_MS_LR I_MS_LR = squeeze(LRMS_tmp); LMS_tmp = lms_multiExm(i, :, :, :); % I_MS I_MS = squeeze(LMS_tmp); PAN_tmp = pan_multiExm(i, :, :, :); % I_PAN I_PAN = squeeze(PAN_tmp); NumIndexes = 3; MatrixResults = zeros(numel(algorithms),NumIndexes); alg = 0; flagQNR = 0; %% Flag QNR/HQNR, 1: QNR otherwise HQNR % for img show location1 = [10 50 190 240]; %default: data6: [10 50 1 60]; data7:[140 180 5 60] location2 = [230 280 240 290]; %default: data6: [190 240 5 60]; data7:[190 235 120 150] sensor = 'GF2'; % disp(sensor) %% load Indexes for WV3_FR % sensor = 'WV3'; Qblocks_size = 32; bicubic = 0;% Interpolator flag_cut_bounds = 1;% Cut Final Image dim_cut = 21;% Cut Final Image thvalues = 0;% Threshold values out of dynamic range printEPS = 0;% Print Eps ratio = 4;% Resize Factor L = 11;% Radiometric Resolution clear print %% show I_MS_LR, I_GT, PAN Imgs: % if size(I_MS,3) == 4 % showImage4LR(I_MS_LR,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % else % showImage8_zoomin(I_MS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % end % % %showPan(I_PAN,printEPS,2,flag_cut_bounds,dim_cut); % showPan_zoomin(I_PAN,printEPS,2,flag_cut_bounds,dim_cut, location1, location2); % pause(2);print('-deps', strcat(data_name, num2str(i-1), '_pan', '.eps')) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%% CS-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% %% ====== 1) BT-H Method ====== if ismember('BT-H',algorithms) alg = alg + 1; cd BT-H t2=tic; % I_BT_H = BroveyRegHazeMin(I_MS,I_PAN,ratio); % time_BT_H = toc(t2); % fprintf('Elaboration time BT-H: %.2f [sec]\n',time_BT_H); cd .. %%% Quality indexes computation [D_lambda_BT_H, D_S_BT_H, QNRI_BT_H] = indexes_evaluation_FS(I_MS,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); MatrixResults(alg,:) = [D_lambda_BT_H,D_S_BT_H,QNRI_BT_H]; % MatrixImage(:,:,:,alg) = I_BT_H; D_lambda_BT_H_multiexm(i) = D_lambda_BT_H; D_S_BT_H_multiexm(i) = D_S_BT_H; QNRI_BT_H_multiexm(i) = QNRI_BT_H; % % showImage8_zoomin(I_BT_H,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bth.eps')) end %% ====== 2) BDSD-PC Method ====== if ismember('BDSD-PC',algorithms) alg = alg + 1; cd BDSD t2=tic; I_BDSD_PC = BDSD_PC(I_MS,I_PAN,ratio,sensor); time_BDSD_PC = toc(t2); fprintf('Elaboration time BDSD-PC: %.2f [sec]\n',time_BDSD_PC); cd .. [D_lambda_BDSD_PC, D_S_BDSD_PC, QNRI_BDSD_PC] = indexes_evaluation_FS(I_BDSD_PC,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); MatrixResults(alg,:) = [D_lambda_BDSD_PC,D_S_BDSD_PC,QNRI_BDSD_PC]; MatrixImage(:,:,:,alg) = I_BDSD_PC; D_lambda_BDSD_PC_multiexm(i) = D_lambda_BDSD_PC; D_S_BDSD_PC_multiexm(i) = D_S_BDSD_PC; QNRI_BDSD_PC_multiexm(i) = QNRI_BDSD_PC; % showImage8_zoomin(I_BDSD_PC,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bdsd_pc.eps')) end % %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % %%%%%%%%%%%%% MRA-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% % %% ====== 1) MTF-GLP-HPM-R Method ====== if ismember('MTF-GLP-HPM-R',algorithms) alg = alg + 1; cd GLP t2=tic; I_MTF_GLP_HPM_R = MTF_GLP_HPM_R(I_MS,I_PAN,sensor,ratio); time_MTF_GLP_HPM_R = toc(t2); fprintf('Elaboration time MTF-GLP: %.2f [sec]\n',time_MTF_GLP_HPM_R); cd .. [D_lambda_MTF_GLP_HPM_R, D_S_MTF_GLP_HPM_R, QNRI_MTF_GLP_HPM_R] = indexes_evaluation_FS(I_MTF_GLP_HPM_R,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); MatrixResults(alg,:) = [D_lambda_MTF_GLP_HPM_R,D_S_MTF_GLP_HPM_R,QNRI_MTF_GLP_HPM_R]; MatrixImage(:,:,:,alg) = I_MTF_GLP_HPM_R; D_lambda_MTF_GLP_HPM_R_multiexm(i) = D_lambda_MTF_GLP_HPM_R; D_S_MTF_GLP_HPM_R_multiexm(i) = D_S_MTF_GLP_HPM_R; QNRI_MTF_GLP_HPM_R_multiexm(i) = QNRI_MTF_GLP_HPM_R; % showImage8_zoomin(I_MTF_GLP_HPM_R,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglp_hpm_r.eps')) end %% ====== 2) MTF-GLP-FS Method ====== if ismember('MTF-GLP-FS',algorithms) alg = alg + 1; cd GLP t2=tic; I_MTF_GLP_FS = MTF_GLP_FS(I_MS,I_PAN,sensor,ratio); time_MTF_GLP_FS = toc(t2); fprintf('Elaboration time MTF-GLP-FS: %.2f [sec]\n',time_MTF_GLP_FS); cd .. [D_lambda_MTF_GLP_FS, D_S_MTF_GLP_FS, QNRI_MTF_GLP_FS] = indexes_evaluation_FS(I_MTF_GLP_FS,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); MatrixResults(alg,:) = [D_lambda_MTF_GLP_FS,D_S_MTF_GLP_FS,QNRI_MTF_GLP_FS]; MatrixImage(:,:,:,alg) = I_MTF_GLP_FS; D_lambda_MTF_GLP_FS_multiexm(i) = D_lambda_MTF_GLP_FS; D_S_MTF_GLP_FS_multiexm(i) = D_S_MTF_GLP_FS; QNRI_MTF_GLP_FS_multiexm(i) = QNRI_MTF_GLP_FS; % showImage8_zoomin(I_MTF_GLP_FS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglpfs.eps')) end %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%% VO-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% %% ====== 1) TV Method ====== if ismember('TV',algorithms) alg = alg + 1; %%%%%%%%%%%%%%%%%%%%%%%%%% Parameters setting %%%%%%%%%%%%%%%%%%%%%%%%%%%%% switch sensor case 'IKONOS' w=[0.1091 0.2127 0.2928 0.3854]; c = 8; alpha=1.064; maxiter=10; lambda = 0.47106; case {'GeoEye1','WV4'} w=[0.1552, 0.3959, 0.2902, 0.1587]; c = 8; alpha=0.75; maxiter=50; lambda = 157.8954; case 'WV3' w=[0.0657 0.1012 0.1537 0.1473 0.1245 0.1545 0.1338 0.1192]; c = 8; alpha=0.75; maxiter=50; lambda = 1.0000e-03; end cd TV t2 = tic; I_TV = TV_pansharpen(I_MS_LR,I_PAN,alpha,lambda,c,maxiter,w); time_TV = toc(t2); fprintf('Elaboration time TV: %.2f [sec]\n',time_TV); cd .. [D_lambda_TV, D_S_TV, QNRI_TV] = indexes_evaluation_FS(I_TV,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); MatrixResults(alg,:) = [D_lambda_TV,D_S_TV,QNRI_TV]; MatrixImage(:,:,:,alg) = I_TV; D_lambda_TV_multiexm(i) = D_lambda_TV; D_S_TV_multiexm(i) = D_S_TV; QNRI_TV_multiexm(i) = QNRI_TV; % % showImage8_zoomin(I_TV,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_tv.eps')) end %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %%%%%%%%%%%%% DL-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% % %% ====== 1) PNN Method ====== % if ismember('PNN',algorithms) % % load(strcat('2_DL_Result/WV3_Full/PNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PNN % I_pnn = double(sr); % % % alg = alg + 1; % [D_lambda_pnn, D_S_pnn, QNRI_pnn] = indexes_evaluation_FS(I_pnn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); % MatrixResults(alg,:) = [D_lambda_pnn,D_S_pnn,QNRI_pnn]; % MatrixImage(:,:,:,alg) = I_pnn; % % D_lambda_pnn_multiexm(i) = D_lambda_pnn; % D_S_pnn_multiexm(i) = D_S_pnn; % QNRI_pnn_multiexm(i) = QNRI_pnn; %% %% showImage8_zoomin(I_pnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); %% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pnn.eps')) % end % % %% ====== 2) PanNet Method ====== % % if ismember('PanNet',algorithms) % load(strcat('2_DL_Result/WV3_Full/PanNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PanNet % I_pannet = double(sr); % % alg = alg + 1; % [D_lambda_pannet, D_S_pannet, QNRI_pannet] = indexes_evaluation_FS(I_pannet,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); % MatrixResults(alg,:) = [D_lambda_pannet,D_S_pannet,QNRI_pannet]; % MatrixImage(:,:,:,alg) = I_pannet; % % % D_lambda_pannet_multiexm(i) = D_lambda_pannet; % D_S_pannet_multiexm(i) = D_S_pannet; % QNRI_pannet_multiexm(i) = QNRI_pannet; %% % showImage8_zoomin(I_pannet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pannet.eps')) % end % % % %%% ====== 3) DiCNN Method ====== % if ismember('DiCNN',algorithms) % load(strcat('2_DL_Result/WV3_Full/DiCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for DiCNN % I_dicnn = double(sr); % % % alg = alg + 1; % [D_lambda_dicnn, D_S_dicnn, QNRI_dicnn] = indexes_evaluation_FS(I_dicnn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); % MatrixResults(alg,:) = [D_lambda_dicnn,D_S_dicnn,QNRI_dicnn]; % MatrixImage(:,:,:,alg) = I_dicnn; % % D_lambda_dicnn_multiexm(i) = D_lambda_dicnn; % D_S_dicnn_multiexm(i) = D_S_dicnn; % QNRI_dicnn_multiexm(i) = QNRI_dicnn; %% % showImage8_zoomin(I_dicnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_dicnn.eps')) % end % %% ====== 4) MSDCNN Method ====== % %if ismember('MSDCNN',algorithms) % load(strcat('2_DL_Result/WV3_Full/MSDCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for MSDCNN % I_msdcnn = double(sr); % % % alg = alg + 1; % [D_lambda_msdcnn, D_S_msdcnn, QNRI_msdcnn] = indexes_evaluation_FS(I_msdcnn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); % MatrixResults(alg,:) = [D_lambda_msdcnn,D_S_msdcnn,QNRI_msdcnn]; % MatrixImage(:,:,:,alg) = I_msdcnn; % % D_lambda_msdcnn_multiexm(i) = D_lambda_msdcnn; % D_S_msdcnn_multiexm(i) = D_S_msdcnn; % QNRI_msdcnn_multiexm(i) = QNRI_msdcnn; % showImage8_zoomin(I_msdcnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_msdcnn.eps')) %end % %% ====== 5) BDPN Method ====== % % % if ismember('BDPN',algorithms) % load(strcat('2_DL_Result/WV3_Full/BDPN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for BDPN % I_bdpn = double(sr); % % % alg = alg + 1; % [D_lambda_bdpn, D_S_bdpn, QNRI_bdpn] = indexes_evaluation_FS(I_bdpn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); % MatrixResults(alg,:) = [D_lambda_bdpn,D_S_bdpn,QNRI_bdpn]; % MatrixImage(:,:,:,alg) = I_bdpn; % % D_lambda_bdpn_multiexm(i) = D_lambda_bdpn; % D_S_bdpn_multiexm(i) = D_S_bdpn; % QNRI_bdpn_multiexm(i) = QNRI_bdpn; % % % showImage8_zoomin(I_bdpn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_bdpn.eps')) % end % %% ====== 6) FusionNet Method ====== % if ismember('FusionNet',algorithms) % file_fusionnet = 'fusionnet_wv3_rs'; % load(strcat('2_DL_Result/WV3_Full/FusionNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for FusionNet % I_fusionnet = double(sr); % % % alg = alg + 1; % [D_lambda_fusionnet, D_S_fusionnet, QNRI_fusionnet] = indexes_evaluation_FS(I_fusionnet,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); % MatrixResults(alg,:) = [D_lambda_fusionnet,D_S_fusionnet,QNRI_fusionnet]; % MatrixImage(:,:,:,alg) = I_fusionnet; % % D_lambda_fusionnet_multiexm(i) = D_lambda_fusionnet; % D_S_fusionnet_multiexm(i) = D_S_fusionnet; % QNRI_fusionnet_multiexm(i) = QNRI_fusionnet; % % showImage8_zoomin(I_fusionnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_fusionnet.eps')) % end % %% ====== 7) LAGConv Method ====== % % if ismember('LAGConv',algorithms) % load(strcat('2_DL_Result/WV3_Full/LAGConv/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for LAGConv % I_lagnet = double(sr); % alg = alg + 1; % [D_lambda_lagnet, D_S_lagnet, QNRI_lagnet] = indexes_evaluation_FS(I_lagnet,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR); % MatrixResults(alg,:) = [D_lambda_lagnet,D_S_lagnet,QNRI_lagnet]; % MatrixImage(:,:,:,alg) = I_lagnet; % % D_lambda_lagnet_multiexm(i) = D_lambda_lagnet; % D_S_lagnet_multiexm(i) = D_S_lagnet; % QNRI_lagnet_multiexm(i) = QNRI_lagnet; %% % showImage8_zoomin(I_lagnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2); % pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_lagnet.eps')) % end end %% Print in LATEX % if flagQNR == 1 % matrix2latex(MatrixResults,'FR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'DL'},{'DS'},{'QNR'}],'alignment','c','format', '%.4f'); % else % matrix2latex(MatrixResults,'FR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'DL'},{'DS'},{'HQNR'}],'alignment','c','format', '%.4f'); % end %% View All if size(I_MS,3) == 4 vect_index_RGB = [3,2,1]; else vect_index_RGB = [5,3,2]; end titleImages = algorithms; % figure, showImagesAll(MatrixImage,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0); % BT_H: average D_lambda_avg avg_D_lambda_BT_H_multiexm = mean(D_lambda_BT_H_multiexm); std_D_lambda_BT_H_multiexm = std(D_lambda_BT_H_multiexm); avg_D_S_BT_H_multiexm = mean(D_S_BT_H_multiexm); std_D_S_BT_H_multiexm = std(D_S_BT_H_multiexm); avg_QNRI_BT_H_multiexm = mean(QNRI_BT_H_multiexm); std_QNRI_BT_H_multiexm = std(QNRI_BT_H_multiexm); Avg_MatrixResults(1,:) = [avg_D_lambda_BT_H_multiexm, std_D_lambda_BT_H_multiexm, ... avg_D_S_BT_H_multiexm, std_D_S_BT_H_multiexm, avg_QNRI_BT_H_multiexm, std_QNRI_BT_H_multiexm ]; % BDSD: average D_lambda_avg % avg_D_lambda_BDSD_PC_multiexm = mean(D_lambda_BDSD_PC_multiexm); % std_D_lambda_BDSD_PC_multiexm = std(D_lambda_BDSD_PC_multiexm); % avg_D_S_BDSD_PC_multiexm = mean(D_S_BDSD_PC_multiexm); % std_D_S_BDSD_PC_multiexm = std(D_S_BDSD_PC_multiexm); % avg_QNRI_BDSD_PC_multiexm = mean(QNRI_BDSD_PC_multiexm); % std_QNRI_BDSD_PC_multiexm = std(QNRI_BDSD_PC_multiexm); % Avg_MatrixResults(2,:) = [avg_D_lambda_BDSD_PC_multiexm, std_D_lambda_BDSD_PC_multiexm, ... % avg_D_S_BDSD_PC_multiexm, std_D_S_BDSD_PC_multiexm, avg_QNRI_BDSD_PC_multiexm, std_QNRI_BDSD_PC_multiexm % ]; % % MTF_GLP: average D_lambda_avg % avg_D_lambda_MTF_GLP_HPM_R_multiexm = mean(D_lambda_MTF_GLP_HPM_R_multiexm); % std_D_lambda_MTF_GLP_HPM_R_multiexm = std(D_lambda_MTF_GLP_HPM_R_multiexm); % avg_D_S_MTF_GLP_HPM_R_multiexm = mean(D_S_MTF_GLP_HPM_R_multiexm); % std_D_S_MTF_GLP_HPM_R_multiexm = std(D_S_MTF_GLP_HPM_R_multiexm); % avg_QNRI_MTF_GLP_HPM_R_multiexm = mean(QNRI_MTF_GLP_HPM_R_multiexm); % std_QNRI_MTF_GLP_HPM_R_multiexm = std(QNRI_MTF_GLP_HPM_R_multiexm); % Avg_MatrixResults(3,:) = [avg_D_lambda_MTF_GLP_HPM_R_multiexm, std_D_lambda_MTF_GLP_HPM_R_multiexm, ... % avg_D_S_MTF_GLP_HPM_R_multiexm, std_D_S_MTF_GLP_HPM_R_multiexm, avg_QNRI_MTF_GLP_HPM_R_multiexm, std_QNRI_MTF_GLP_HPM_R_multiexm % ]; % % MTF_GLP_FS: average D_lambda_avg % avg_D_lambda_MTF_GLP_FS_multiexm = mean(D_lambda_MTF_GLP_FS_multiexm); % std_D_lambda_MTF_GLP_FS_multiexm = std(D_lambda_MTF_GLP_FS_multiexm); % avg_D_S_MTF_GLP_FS_multiexm = mean(D_S_MTF_GLP_FS_multiexm); % std_D_S_MTF_GLP_FS_multiexm = std(D_S_MTF_GLP_FS_multiexm); % avg_QNRI_MTF_GLP_FS_multiexm = mean(QNRI_MTF_GLP_FS_multiexm); % std_QNRI_MTF_GLP_FS_multiexm = std(QNRI_MTF_GLP_FS_multiexm); % Avg_MatrixResults(4,:) = [avg_D_lambda_MTF_GLP_FS_multiexm, std_D_lambda_MTF_GLP_FS_multiexm, ... % avg_D_S_MTF_GLP_FS_multiexm, std_D_S_MTF_GLP_FS_multiexm, avg_QNRI_MTF_GLP_FS_multiexm, std_QNRI_MTF_GLP_FS_multiexm % ]; % % TV: average D_lambda_avg % avg_D_lambda_TV_multiexm = mean(D_lambda_TV_multiexm); % std_D_lambda_TV_multiexm = std(D_lambda_TV_multiexm); % avg_D_S_TV_multiexm = mean(D_S_TV_multiexm); % std_D_S_TV_multiexm = std(D_S_TV_multiexm); % avg_QNRI_TV_multiexm = mean(QNRI_TV_multiexm); % std_QNRI_TV_multiexm = std(QNRI_TV_multiexm); % Avg_MatrixResults(5,:) = [avg_D_lambda_TV_multiexm, std_D_lambda_TV_multiexm, ... % avg_D_S_TV_multiexm, std_D_S_TV_multiexm, avg_QNRI_TV_multiexm, std_QNRI_TV_multiexm]; %% pnn: average D_lambda_avg %avg_D_lambda_pnn_multiexm = mean(D_lambda_pnn_multiexm); %std_D_lambda_pnn_multiexm = std(D_lambda_pnn_multiexm); % %avg_D_S_pnn_multiexm = mean(D_S_pnn_multiexm); %std_D_S_pnn_multiexm = std(D_S_pnn_multiexm); % %avg_QNRI_pnn_multiexm = mean(QNRI_pnn_multiexm); %std_QNRI_pnn_multiexm = std(QNRI_pnn_multiexm); % % %Avg_MatrixResults(6,:) = [avg_D_lambda_pnn_multiexm, std_D_lambda_pnn_multiexm, ... % avg_D_S_pnn_multiexm, std_D_S_pnn_multiexm, avg_QNRI_pnn_multiexm, std_QNRI_pnn_multiexm % ]; %% pannet: average D_lambda_avg %avg_D_lambda_pannet_multiexm = mean(D_lambda_pannet_multiexm); %std_D_lambda_pannet_multiexm = std(D_lambda_pannet_multiexm); % % %avg_D_S_pannet_multiexm = mean(D_S_pannet_multiexm); %std_D_S_pannet_multiexm = std(D_S_pannet_multiexm); % %avg_QNRI_pannet_multiexm = mean(QNRI_pannet_multiexm); %std_QNRI_pannet_multiexm = std(QNRI_pannet_multiexm); % % %Avg_MatrixResults(7,:) = [avg_D_lambda_pannet_multiexm, std_D_lambda_pannet_multiexm, ... % avg_D_S_pannet_multiexm, std_D_S_pannet_multiexm, avg_QNRI_pannet_multiexm, std_QNRI_pannet_multiexm % ]; % % % %% dicnn: average D_lambda_avg %avg_D_lambda_dicnn_multiexm = mean(D_lambda_dicnn_multiexm); %std_D_lambda_dicnn_multiexm = std(D_lambda_dicnn_multiexm); % %avg_D_S_dicnn_multiexm = mean(D_S_dicnn_multiexm); %std_D_S_dicnn_multiexm = std(D_S_dicnn_multiexm); % %avg_QNRI_dicnn_multiexm = mean(QNRI_dicnn_multiexm); %std_QNRI_dicnn_multiexm = std(QNRI_dicnn_multiexm); % %Avg_MatrixResults(8,:) = [avg_D_lambda_dicnn_multiexm, std_D_lambda_dicnn_multiexm, ... % avg_D_S_dicnn_multiexm, std_D_S_dicnn_multiexm, avg_QNRI_dicnn_multiexm, std_QNRI_dicnn_multiexm % ]; %% msdcnn: average D_lambda_avg %avg_D_lambda_msdcnn_multiexm = mean(D_lambda_msdcnn_multiexm); %std_D_lambda_msdcnn_multiexm = std(D_lambda_msdcnn_multiexm); % % %avg_D_S_msdcnn_multiexm = mean(D_S_msdcnn_multiexm); %std_D_S_msdcnn_multiexm = std(D_S_msdcnn_multiexm); % %avg_QNRI_msdcnn_multiexm = mean(QNRI_msdcnn_multiexm); %std_QNRI_msdcnn_multiexm = std(QNRI_msdcnn_multiexm); % %Avg_MatrixResults(9,:) = [avg_D_lambda_msdcnn_multiexm, std_D_lambda_msdcnn_multiexm, ... % avg_D_S_msdcnn_multiexm, std_D_S_msdcnn_multiexm, avg_QNRI_msdcnn_multiexm, std_QNRI_msdcnn_multiexm % ]; %% bdpn: average D_lambda_avg %avg_D_lambda_bdpn_multiexm = mean(D_lambda_bdpn_multiexm); %std_D_lambda_bdpn_multiexm = std(D_lambda_bdpn_multiexm); % % %avg_D_S_bdpn_multiexm = mean(D_S_bdpn_multiexm); %std_D_S_bdpn_multiexm = std(D_S_bdpn_multiexm); % %avg_QNRI_bdpn_multiexm = mean(QNRI_bdpn_multiexm); %std_QNRI_bdpn_multiexm = std(QNRI_bdpn_multiexm); % %Avg_MatrixResults(10,:) = [avg_D_lambda_bdpn_multiexm, std_D_lambda_bdpn_multiexm, ... % avg_D_S_bdpn_multiexm, std_D_S_bdpn_multiexm, avg_QNRI_bdpn_multiexm, std_QNRI_bdpn_multiexm % ]; % % % % %% %% fusionnet: average D_lambda_avg %avg_D_lambda_fusionnet_multiexm = mean(D_lambda_fusionnet_multiexm); %std_D_lambda_fusionnet_multiexm = std(D_lambda_fusionnet_multiexm); % %avg_D_S_fusionnet_multiexm = mean(D_S_fusionnet_multiexm); %std_D_S_fusionnet_multiexm = std(D_S_fusionnet_multiexm); % %avg_QNRI_fusionnet_multiexm = mean(QNRI_fusionnet_multiexm); %std_QNRI_fusionnet_multiexm = std(QNRI_fusionnet_multiexm); % %Avg_MatrixResults(11,:) = [avg_D_lambda_fusionnet_multiexm, std_D_lambda_fusionnet_multiexm, ... % avg_D_S_fusionnet_multiexm, std_D_S_fusionnet_multiexm, avg_QNRI_fusionnet_multiexm, std_QNRI_fusionnet_multiexm % ]; %% lagnet: average D_lambda_avg %avg_D_lambda_lagnet_multiexm = mean(D_lambda_lagnet_multiexm); %std_D_lambda_lagnet_multiexm = std(D_lambda_lagnet_multiexm); % %avg_D_S_lagnet_multiexm = mean(D_S_lagnet_multiexm); %std_D_S_lagnet_multiexm = std(D_S_lagnet_multiexm); % %avg_QNRI_lagnet_multiexm = mean(QNRI_lagnet_multiexm); %std_QNRI_lagnet_multiexm = std(QNRI_lagnet_multiexm); % % %Avg_MatrixResults(12,:) = [avg_D_lambda_lagnet_multiexm, std_D_lambda_lagnet_multiexm, ... % avg_D_S_lagnet_multiexm, std_D_S_lagnet_multiexm, avg_QNRI_lagnet_multiexm, std_QNRI_lagnet_multiexm % ]; matrix2latex(Avg_MatrixResults,'Avg_RR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'D_lambda'}, {'D_l-std'}, {'D_S'}, {'D_S-std'}, {'QNRI'}, {'QNRI-std'}],'alignment','c','format', '%.4f'); fprintf('\n') disp('#######################################################') disp(['Display the performance for:', num2str(1:i)]) disp('#######################################################') disp(' |===D_lambda_avg===|=====D_s=====|======QNR=======') Avg_MatrixResults
281677160/openwrt-package
86,169
luci-app-passwall/root/usr/share/passwall/nftables.sh
#!/bin/bash DIR="$(cd "$(dirname "$0")" && pwd)" MY_PATH=$DIR/nftables.sh NFTABLE_NAME="inet passwall" NFTSET_LOCAL="passwall_local" NFTSET_LAN="passwall_lan" NFTSET_VPS="passwall_vps" NFTSET_SHUNT="passwall_shunt" NFTSET_GFW="passwall_gfw" NFTSET_CHN="passwall_chn" NFTSET_BLACK="passwall_black" NFTSET_WHITE="passwall_white" NFTSET_BLOCK="passwall_block" NFTSET_LOCAL6="passwall_local6" NFTSET_LAN6="passwall_lan6" NFTSET_VPS6="passwall_vps6" NFTSET_SHUNT6="passwall_shunt6" NFTSET_GFW6="passwall_gfw6" NFTSET_CHN6="passwall_chn6" NFTSET_BLACK6="passwall_black6" NFTSET_WHITE6="passwall_white6" NFTSET_BLOCK6="passwall_block6" FORCE_INDEX=0 USE_SHUNT_TCP=0 USE_SHUNT_UDP=0 . /lib/functions/network.sh FWI=$(uci -q get firewall.passwall.path 2>/dev/null) FAKE_IP="198.18.0.0/15" FAKE_IP_6="fc00::/18" factor() { local ports="$1" if [ -z "$1" ] || [ -z "$2" ] || [ "$ports" = "1:65535" ]; then echo "" # acl mac address elif echo "$1" | grep -qE '([A-Fa-f0-9]{2}:){5}[A-Fa-f0-9]{2}'; then echo "$2 {$1}" else ports=$(echo "$ports" | tr -d ' ' | sed 's/:/-/g' | tr ',' '\n' | awk '!a[$0]++' | grep -v '^$') [ -z "$ports" ] && { echo ""; return; } if echo "$ports" | grep -q '^1-65535$'; then echo "" return fi local port local port_list="" for port in $ports; do port_list="${port_list},$port" done port_list="${port_list#,}" echo "$2 {$port_list}" fi } insert_rule_before() { [ $# -ge 4 ] || { return 1 } local table_name="${1}"; shift local chain_name="${1}"; shift local keyword="${1}"; shift local rule="${1}"; shift local default_index="${1}"; shift default_index=${default_index:-0} local _index=$(nft -a list chain $table_name $chain_name 2>/dev/null | grep "$keyword" | awk -F '# handle ' '{print$2}' | head -n 1 | awk '{print $1}') if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then nft "add rule $table_name $chain_name $rule" else if [ -z "${_index}" ]; then _index=${default_index} fi nft "insert rule $table_name $chain_name position $_index $rule" fi } insert_rule_after() { [ $# -ge 4 ] || { return 1 } local table_name="${1}"; shift local chain_name="${1}"; shift local keyword="${1}"; shift local rule="${1}"; shift local default_index="${1}"; shift default_index=${default_index:-0} local _index=$(nft -a list chain $table_name $chain_name 2>/dev/null | grep "$keyword" | awk -F '# handle ' '{print$2}' | head -n 1 | awk '{print $1}') if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then nft "add rule $table_name $chain_name $rule" else if [ -n "${_index}" ]; then _index=$((_index + 1)) else _index=${default_index} fi nft "insert rule $table_name $chain_name position $_index $rule" fi } RULE_LAST_INDEX() { [ $# -ge 3 ] || { echolog "索引列举方式不正确(nftables),终止执行!" return 1 } local table_name="${1}"; shift local chain_name="${1}"; shift local keyword="${1}"; shift local default="${1:-0}"; shift local _index=$(nft -a list chain $table_name $chain_name 2>/dev/null | grep "$keyword" | awk -F '# handle ' '{print$2}' | head -n 1 | awk '{print $1}') echo "${_index:-${default}}" } REDIRECT() { local s="counter redirect" [ -n "$1" ] && { local s="$s to :$1" [ "$2" == "MARK" ] && s="counter meta mark set $1" [ "$2" == "TPROXY" ] && { s="counter meta mark 1 tproxy to :$1" } [ "$2" == "TPROXY4" ] && { s="counter meta mark 1 tproxy ip to :$1" } [ "$2" == "TPROXY6" ] && { s="counter meta mark 1 tproxy ip6 to :$1" } } echo $s } destroy_nftset() { for i in "$@"; do nft flush set $NFTABLE_NAME $i 2>/dev/null nft delete set $NFTABLE_NAME $i 2>/dev/null done } gen_nft_tables() { if ! nft list tables | grep -q "^table inet passwall$"; then nft -f - <<-EOF table $NFTABLE_NAME { chain dstnat { type nat hook prerouting priority dstnat - 1; policy accept; } chain mangle_prerouting { type filter hook prerouting priority mangle - 1; policy accept; } chain mangle_output { type route hook output priority mangle - 1; policy accept; } chain nat_output { type nat hook output priority -1; policy accept; } } EOF fi } insert_nftset() { local nftset_name="${1}"; shift local timeout_argument="${1}"; shift local default_timeout_argument="365d" [ -n "${1}" ] && { local nftset_elements case "$timeout_argument" in "-1") nftset_elements=$(echo -e $@ | sed 's/\s/, /g') ;; "0") nftset_elements=$(echo -e $@ | sed "s/\s/ timeout $default_timeout_argument, /g" | sed "s/$/ timeout $default_timeout_argument/") ;; *) nftset_elements=$(echo -e $@ | sed "s/\s/ timeout $timeout_argument, /g" | sed "s/$/ timeout $timeout_argument/") ;; esac mkdir -p $TMP_PATH2/nftset cat > "$TMP_PATH2/nftset/$nftset_name" <<-EOF define $nftset_name = {$nftset_elements} add element $NFTABLE_NAME $nftset_name \$$nftset_name EOF nft -f "$TMP_PATH2/nftset/$nftset_name" rm -rf "$TMP_PATH2/nftset" } } gen_nftset() { local nftset_name="${1}"; shift local ip_type="${1}"; shift # 0 - don't set defalut timeout local timeout_argument_set="${1}"; shift # 0 - don't let element timeout(365 days) when set's timeout parameters be seted # -1 - follow the set's timeout parameters local timeout_argument_element="${1}"; shift nft "list set $NFTABLE_NAME $nftset_name" &>/dev/null if [ $? -ne 0 ]; then if [ "$timeout_argument_set" == "0" ]; then nft "add set $NFTABLE_NAME $nftset_name { type $ip_type; flags interval, timeout; auto-merge; }" else nft "add set $NFTABLE_NAME $nftset_name { type $ip_type; flags interval, timeout; timeout $timeout_argument_set; gc-interval $timeout_argument_set; auto-merge; }" fi fi [ -n "${1}" ] && insert_nftset $nftset_name $timeout_argument_element $@ } get_jump_ipt() { case "$1" in direct) echo "mark != 1 counter return" ;; proxy) if [ -n "$2" ] && [ -n "$(echo $2 | grep "^counter")" ]; then echo "$2" else echo "$(REDIRECT $2 $3)" fi ;; esac } gen_lanlist() { cat $RULES_PATH/lanlist_ipv4 | tr -s '\n' | grep -v "^#" } gen_lanlist_6() { cat $RULES_PATH/lanlist_ipv6 | tr -s '\n' | grep -v "^#" } get_wan_ip() { local NET_IF local NET_ADDR network_flush_cache network_find_wan NET_IF network_get_ipaddr NET_ADDR "${NET_IF}" echo $NET_ADDR } get_wan6_ip() { local NET_IF local NET_ADDR network_flush_cache network_find_wan6 NET_IF network_get_ipaddr6 NET_ADDR "${NET_IF}" echo $NET_ADDR } load_acl() { ([ "$ENABLED_ACLS" == 1 ] || ([ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ])) && echolog " - 访问控制:" [ "$ENABLED_ACLS" == 1 ] && { acl_app for sid in $(ls -F ${TMP_ACL_PATH} | grep '/$' | awk -F '/' '{print $1}' | grep -v 'default'); do eval $(uci -q show "${CONFIG}.${sid}" | cut -d'.' -sf 3-) tcp_no_redir_ports=${tcp_no_redir_ports:-default} udp_no_redir_ports=${udp_no_redir_ports:-default} use_global_config=${use_global_config:-0} tcp_proxy_drop_ports=${tcp_proxy_drop_ports:-default} udp_proxy_drop_ports=${udp_proxy_drop_ports:-default} tcp_redir_ports=${tcp_redir_ports:-default} udp_redir_ports=${udp_redir_ports:-default} use_direct_list=${use_direct_list:-1} use_proxy_list=${use_proxy_list:-1} use_block_list=${use_block_list:-1} use_gfw_list=${use_gfw_list:-1} chn_list=${chn_list:-direct} tcp_proxy_mode=${tcp_proxy_mode:-proxy} udp_proxy_mode=${udp_proxy_mode:-proxy} [ "$tcp_no_redir_ports" = "default" ] && tcp_no_redir_ports=$TCP_NO_REDIR_PORTS [ "$udp_no_redir_ports" = "default" ] && udp_no_redir_ports=$UDP_NO_REDIR_PORTS [ "$tcp_proxy_drop_ports" = "default" ] && tcp_proxy_drop_ports=$TCP_PROXY_DROP_PORTS [ "$udp_proxy_drop_ports" = "default" ] && udp_proxy_drop_ports=$UDP_PROXY_DROP_PORTS [ "$tcp_redir_ports" = "default" ] && tcp_redir_ports=$TCP_REDIR_PORTS [ "$udp_redir_ports" = "default" ] && udp_redir_ports=$UDP_REDIR_PORTS [ -n "$(get_cache_var "ACL_${sid}_tcp_node")" ] && tcp_node=$(get_cache_var "ACL_${sid}_tcp_node") [ -n "$(get_cache_var "ACL_${sid}_tcp_redir_port")" ] && tcp_port=$(get_cache_var "ACL_${sid}_tcp_redir_port") [ -n "$(get_cache_var "ACL_${sid}_udp_node")" ] && udp_node=$(get_cache_var "ACL_${sid}_udp_node") [ -n "$(get_cache_var "ACL_${sid}_udp_redir_port")" ] && udp_port=$(get_cache_var "ACL_${sid}_udp_redir_port") [ -n "$(get_cache_var "ACL_${sid}_dns_port")" ] && dns_redirect_port=$(get_cache_var "ACL_${sid}_dns_port") [ -n "$tcp_node" ] && tcp_node_remark=$(config_n_get $tcp_node remarks) [ -n "$udp_node" ] && udp_node_remark=$(config_n_get $udp_node remarks) use_shunt_tcp=0 use_shunt_udp=0 [ -n "$tcp_node" ] && [ "$(config_n_get $tcp_node protocol)" = "_shunt" ] && use_shunt_tcp=1 [ -n "$udp_node" ] && [ "$(config_n_get $udp_node protocol)" = "_shunt" ] && use_shunt_udp=1 [ "${use_global_config}" = "1" ] && { tcp_node_remark=$(config_n_get $TCP_NODE remarks) udp_node_remark=$(config_n_get $UDP_NODE remarks) use_direct_list=${USE_DIRECT_LIST} use_proxy_list=${USE_PROXY_LIST} use_block_list=${USE_BLOCK_LIST} use_gfw_list=${USE_GFW_LIST} chn_list=${CHN_LIST} tcp_proxy_mode=${TCP_PROXY_MODE} udp_proxy_mode=${UDP_PROXY_MODE} use_shunt_tcp=${USE_SHUNT_TCP} use_shunt_udp=${USE_SHUNT_UDP} dns_redirect_port=${DNS_REDIRECT_PORT} black_set_name=${NFTSET_BLACK} black6_set_name=${NFTSET_BLACK6} gfw_set_name=${NFTSET_GFW} gfw6_set_name=${NFTSET_GFW6} shunt_set_name=${NFTSET_SHUNT} shunt6_set_name=${NFTSET_SHUNT6} } _acl_list=${TMP_ACL_PATH}/${sid}/source_list for i in $(cat $_acl_list); do local _ipt_source _ipv4 local msg if [ -n "${interface}" ]; then . /lib/functions/network.sh local gateway device network_get_gateway gateway "${interface}" network_get_device device "${interface}" [ -z "${device}" ] && device="${interface}" _ipt_source="iifname ${device} " msg="源接口【${device}】," else msg="源接口【所有】," fi if [ -n "$(echo ${i} | grep '^iprange:')" ]; then _iprange=$(echo ${i} | sed 's#iprange:##g') _ipt_source=$(factor ${_iprange} "${_ipt_source}ip saddr") msg="${msg}IP range【${_iprange}】," _ipv4="1" unset _iprange elif [ -n "$(echo ${i} | grep '^ipset:')" ]; then _ipset=$(echo ${i} | sed 's#ipset:##g') _ipt_source="${_ipt_source}ip saddr @${_ipset}" msg="${msg}NFTset【${_ipset}】," unset _ipset elif [ -n "$(echo ${i} | grep '^ip:')" ]; then _ip=$(echo ${i} | sed 's#ip:##g') _ipt_source=$(factor ${_ip} "${_ipt_source}ip saddr") msg="${msg}IP【${_ip}】," _ipv4="1" unset _ip elif [ -n "$(echo ${i} | grep '^mac:')" ]; then _mac=$(echo ${i} | sed 's#mac:##g') _ipt_source=$(factor ${_mac} "${_ipt_source}ether saddr") msg="${msg}MAC【${_mac}】," unset _mac elif [ -n "$(echo ${i} | grep '^any')" ]; then msg="${msg}所有设备," else continue fi msg="【$remarks】,${msg}" [ "$tcp_no_redir_ports" != "disable" ] && { if ! has_1_65535 "$tcp_no_redir_ports"; then nft "add rule $NFTABLE_NAME $nft_prerouting_chain ${_ipt_source} ip protocol tcp $(factor $tcp_no_redir_ports "tcp dport") counter return comment \"$remarks\"" [ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ${_ipt_source} meta l4proto tcp $(factor $tcp_no_redir_ports "tcp dport") counter return comment \"$remarks\"" echolog " - ${msg}不代理 TCP 端口[${tcp_no_redir_ports}]" else #结束时会return,无需加多余的规则。 unset tcp_port echolog " - ${msg}不代理所有 TCP 端口" fi } [ "$udp_no_redir_ports" != "disable" ] && { if ! has_1_65535 "$udp_no_redir_ports"; then nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_no_redir_ports "udp dport") counter return comment \"$remarks\"" [ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_no_redir_ports "udp dport") counter return comment \"$remarks\"" 2>/dev/null echolog " - ${msg}不代理 UDP 端口[${udp_no_redir_ports}]" else #结束时会return,无需加多余的规则。 unset udp_port echolog " - ${msg}不代理所有 UDP 端口" fi } local dns_redirect [ $(config_t_get global dns_redirect "1") = "1" ] && dns_redirect=53 if ([ -n "$tcp_port" ] && [ -n "${tcp_proxy_mode}" ]) || ([ -n "$udp_port" ] && [ -n "${udp_proxy_mode}" ]); then [ "${use_proxy_list}" = "1" ] && { [ "${use_global_config}" = "0" ] && { black_set_name="passwall_${sid}_black" black6_set_name="passwall_${sid}_black6" gen_nftset $black_set_name ipv4_addr "2d" 0 gen_nftset $black6_set_name ipv6_addr "2d" 0 } } [ "${use_gfw_list}" = "1" ] && { [ "${use_global_config}" = "0" ] && { gfw_set_name="passwall_${sid}_gfw" gfw6_set_name="passwall_${sid}_gfw6" gen_nftset $gfw_set_name ipv4_addr "2d" 0 gen_nftset $gfw6_set_name ipv6_addr "2d" 0 } } [ "${use_shunt_tcp}" = "1" -o "${use_shunt_udp}" = "1" ] && { [ "${use_global_config}" = "0" ] && { shunt_set_name="passwall_${sid}_shunt" shunt6_set_name="passwall_${sid}_shunt6" gen_nftset $shunt_set_name ipv4_addr "2d" 0 gen_nftset $shunt6_set_name ipv6_addr "2d" 0 } } [ -n "${dns_redirect_port}" ] && dns_redirect=${dns_redirect_port} else [ -n "${DIRECT_DNSMASQ_PORT}" ] && dns_redirect=${DIRECT_DNSMASQ_PORT} fi if [ -n "${dns_redirect}" ]; then nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} udp dport 53 counter return comment \"$remarks\"" [ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} udp dport 53 counter return comment \"$remarks\"" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp ${_ipt_source} tcp dport 53 counter return comment \"$remarks\"" [ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} tcp dport 53 counter return comment \"$remarks\"" #nft "add rule $NFTABLE_NAME PSW_DNS ip protocol udp ${_ipt_source} udp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\"" #nft "add rule $NFTABLE_NAME PSW_DNS ip protocol tcp ${_ipt_source} tcp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\"" nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto udp ${_ipt_source} udp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\"" nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto tcp ${_ipt_source} tcp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\"" [ -z "$(get_cache_var "ACL_${sid}_tcp_default")" ] && echolog " - ${msg}使用与全局配置不相同节点,已将DNS强制重定向到专用 DNS 服务器。" fi [ -n "$tcp_port" -o -n "$udp_port" ] && { [ "${use_block_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ${_ipt_source} ip daddr @$NFTSET_BLOCK counter drop comment \"$remarks\"" [ "${use_block_list}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ${_ipt_source} ip daddr @$NFTSET_BLOCK counter drop comment \"$remarks\"" [ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ${_ipt_source} ip daddr @$NFTSET_WHITE counter return comment \"$remarks\"" [ "${use_direct_list}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ${_ipt_source} ip daddr @$NFTSET_WHITE counter return comment \"$remarks\"" [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { [ "${use_block_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ${_ipt_source} ip6 daddr @$NFTSET_BLOCK6 counter drop comment \"$remarks\"" [ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ${_ipt_source} ip6 daddr @$NFTSET_WHITE6 counter return comment \"$remarks\"" } [ "$tcp_proxy_drop_ports" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"$remarks\"" 2>/dev/null [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$black6_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$gfw6_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\"" 2>/dev/null [ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$shunt6_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") counter drop comment \"$remarks\"" 2>/dev/null } nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr $FAKE_IP counter drop comment \"$remarks\"" [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$black_set_name counter drop comment \"$remarks\"" [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$gfw_set_name counter drop comment \"$remarks\"" [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\"" [ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$shunt_set_name counter drop comment \"$remarks\"" [ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") counter drop comment \"$remarks\"" echolog " - ${msg}屏蔽代理 TCP 端口[${tcp_proxy_drop_ports}]" } [ "$udp_proxy_drop_ports" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"$remarks\"" 2>/dev/null [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$black6_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$gfw6_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\"" 2>/dev/null [ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$shunt6_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") counter drop comment \"$remarks\"" 2>/dev/null } nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr $FAKE_IP counter drop comment \"$remarks\"" 2>/dev/null [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$black_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$gfw_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\"" 2>/dev/null [ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$shunt_set_name counter drop comment \"$remarks\"" 2>/dev/null [ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") counter drop comment \"$remarks\"" 2>/dev/null echolog " - ${msg}屏蔽代理 UDP 端口[${udp_proxy_drop_ports}]" } } [ -n "$tcp_port" ] && { if [ -n "${tcp_proxy_mode}" ]; then msg2="${msg}使用 TCP 节点[$tcp_node_remark]" if [ -n "${is_tproxy}" ]; then msg2="${msg2}(TPROXY:${tcp_port})" nft_chain="PSW_MANGLE" nft_j="counter jump PSW_RULE" else msg2="${msg2}(REDIRECT:${tcp_port})" nft_chain="PSW_NAT" nft_j="$(REDIRECT $tcp_port)" fi [ "$accept_icmp" = "1" ] && { [ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$NFTSET_WHITE counter return comment \"$remarks\"" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr $FAKE_IP $(REDIRECT) comment \"$remarks\"" [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$black_set_name $(REDIRECT) comment \"$remarks\"" [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$gfw_set_name $(REDIRECT) comment \"$remarks\"" [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list}) comment \"$remarks\"" [ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$shunt_set_name $(REDIRECT) comment \"$remarks\"" [ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} $(REDIRECT) comment \"$remarks\"" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} return comment \"$remarks\"" } [ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { [ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$NFTSET_WHITE6 counter return comment \"$remarks\"" 2>/dev/null nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr $FAKE_IP_6 $(REDIRECT) comment \"$remarks\"" 2>/dev/null [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$black6_set_name $(REDIRECT) comment \"$remarks\"" 2>/dev/null [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$gfw6_set_name $(REDIRECT) comment \"$remarks\"" 2>/dev/null [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list}) comment \"$remarks\"" 2>/dev/null [ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$shunt6_set_name $(REDIRECT) comment \"$remarks\"" 2>/dev/null [ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} $(REDIRECT) comment \"$remarks\"" 2>/dev/null nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} return comment \"$remarks\"" 2>/dev/null } nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} ip daddr $FAKE_IP ${nft_j} comment \"$remarks\"" [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$black_set_name ${nft_j} comment \"$remarks\" " [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$gfw_set_name ${nft_j} comment \"$remarks\" " [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "${nft_j}") comment \"$remarks\" " [ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$shunt_set_name ${nft_j} comment \"$remarks\"" [ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ${nft_j} comment \"$remarks\"" [ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY4) comment \"$remarks\"" [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"$remarks\"" [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$black6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$gfw6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter jump PSW_RULE") comment \"$remarks\" " [ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$shunt6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null [ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY) comment \"$remarks\"" 2>/dev/null } else msg2="${msg}不代理 TCP" fi echolog " - ${msg2}" } nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} counter return comment \"$remarks\"" [ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} counter return comment \"$remarks\"" 2>/dev/null [ -n "$udp_port" ] && { if [ -n "${udp_proxy_mode}" ]; then msg2="${msg}使用 UDP 节点[$udp_node_remark]" msg2="${msg2}(TPROXY:${udp_port})" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} ip daddr $FAKE_IP counter jump PSW_RULE comment \"$remarks\"" [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$black_set_name counter jump PSW_RULE comment \"$remarks\"" [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$gfw_set_name counter jump PSW_RULE comment \"$remarks\"" [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "counter jump PSW_RULE") comment \"$remarks\"" [ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$shunt_set_name counter jump PSW_RULE comment \"$remarks\"" [ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") counter jump PSW_RULE comment \"$remarks\"" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(REDIRECT $udp_port TPROXY4) comment \"$remarks\"" [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"$remarks\"" [ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$black6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null [ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$gfw6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null [ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter jump PSW_RULE") comment \"$remarks\"" 2>/dev/null [ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$shunt6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null [ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(REDIRECT $udp_port TPROXY) comment \"$remarks\"" 2>/dev/null } else msg2="${msg}不代理 UDP" fi echolog " - ${msg2}" } nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} counter return comment \"$remarks\"" [ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} counter return comment \"$remarks\"" 2>/dev/null unset nft_chain nft_j _ipt_source msg msg2 _ipv4 done unset enabled sid remarks sources use_global_config use_direct_list use_proxy_list use_block_list use_gfw_list chn_list tcp_proxy_mode udp_proxy_mode dns_redirect_port tcp_no_redir_ports udp_no_redir_ports tcp_proxy_drop_ports udp_proxy_drop_ports tcp_redir_ports udp_redir_ports tcp_node udp_node interface unset tcp_port udp_port tcp_node_remark udp_node_remark _acl_list use_shunt_tcp use_shunt_udp dns_redirect done } [ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ] && { msg="【默认】," [ "$TCP_NO_REDIR_PORTS" != "disable" ] && { nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return comment \"默认\"" if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]" else unset TCP_PROXY_MODE echolog " - ${msg}不代理所有 TCP 端口" fi } [ "$UDP_NO_REDIR_PORTS" != "disable" ] && { nft "add $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return comment \"默认\"" nft "add $NFTABLE_NAME PSW_MANGLE_V6 counter meta l4proto udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return comment \"默认\"" if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]" else unset UDP_PROXY_MODE echolog " - ${msg}不代理所有 UDP 端口" fi } local DNS_REDIRECT [ $(config_t_get global dns_redirect "1") = "1" ] && DNS_REDIRECT=53 if ([ -n "$TCP_NODE" ] && [ -n "${TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${UDP_PROXY_MODE}" ]); then [ -n "${DNS_REDIRECT_PORT}" ] && DNS_REDIRECT=${DNS_REDIRECT_PORT} else [ -n "${DIRECT_DNSMASQ_PORT}" ] && DNS_REDIRECT=${DIRECT_DNSMASQ_PORT} fi if [ -n "${DNS_REDIRECT}" ]; then nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp udp dport 53 counter return comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp udp dport 53 counter return comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp tcp dport 53 counter return comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp tcp dport 53 counter return comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_DNS ip protocol udp udp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_DNS ip protocol tcp tcp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto udp udp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto tcp tcp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\"" fi [ -n "${TCP_PROXY_MODE}" -o -n "${UDP_PROXY_MODE}" ] && { [ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_BLOCK counter drop comment \"默认\"" [ "${USE_BLOCK_LIST}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_BLOCK counter drop comment \"默认\"" [ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_WHITE counter return comment \"默认\"" [ "${USE_DIRECT_LIST}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_WHITE counter return comment \"默认\"" [ "$PROXY_IPV6" == "1" ] && { [ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_BLOCK6 counter drop comment \"默认\"" [ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_WHITE6 counter return comment \"默认\"" } [ "$TCP_PROXY_DROP_PORTS" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_BLACK6 counter drop comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_GFW6 counter drop comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\"" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_SHUNT6 counter drop comment \"默认\"" [ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop comment \"默认\"" } nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr $FAKE_IP counter drop comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_BLACK counter drop comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_GFW counter drop comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\"" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_SHUNT counter drop comment \"默认\"" [ "${TCP_PROXY_MODE}" != "disable" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop comment \"默认\"" echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]" } [ "$UDP_PROXY_DROP_PORTS" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_BLACK6 counter drop comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_GFW6 counter drop comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\"" [ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_SHUNT6 counter drop comment \"默认\"" [ "${UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop comment \"默认\"" } nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr $FAKE_IP counter drop comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_BLACK counter drop comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_GFW counter drop comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\"" [ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_SHUNT counter drop comment \"默认\"" [ "${UDP_PROXY_MODE}" != "disable" ] && nft "add $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop comment \"默认\"" echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]" } } # 加载TCP默认代理模式 if [ -n "${TCP_PROXY_MODE}" ]; then [ -n "$TCP_NODE" ] && { msg2="${msg}使用 TCP 节点[$(config_n_get $TCP_NODE remarks)]" if [ -n "${is_tproxy}" ]; then msg2="${msg2}(TPROXY:${TCP_REDIR_PORT})" nft_chain="PSW_MANGLE" nft_j="counter jump PSW_RULE" else msg2="${msg2}(REDIRECT:${TCP_REDIR_PORT})" nft_chain="PSW_NAT" nft_j="$(REDIRECT $TCP_REDIR_PORT)" fi [ "$accept_icmp" = "1" ] && { [ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip daddr @$NFTSET_WHITE counter return comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr $FAKE_IP $(REDIRECT) comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_BLACK $(REDIRECT) comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_GFW $(REDIRECT) comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST}) comment \"默认\"" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_SHUNT $(REDIRECT) comment \"默认\"" [ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp $(REDIRECT) comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp return comment \"默认\"" } [ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && { [ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip6 daddr @$NFTSET_WHITE6 counter return comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr $FAKE_IP_6 $(REDIRECT) comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_BLACK6 $(REDIRECT) comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_GFW6 $(REDIRECT) comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST}) comment \"默认\"" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_SHUNT6 $(REDIRECT) comment \"默认\"" [ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 $(REDIRECT) comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 return comment \"默认\"" } nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr $FAKE_IP ${nft_j} comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_BLACK ${nft_j} comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_GFW ${nft_j} comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "${nft_j}") comment \"默认\"" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_SHUNT ${nft_j} comment \"默认\"" [ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j} comment \"默认\"" [ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(REDIRECT $TCP_REDIR_PORT TPROXY4) comment \"默认\"" nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp counter return comment \"默认\"" [ "$PROXY_IPV6" == "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_BLACK6 counter jump PSW_RULE comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_GFW6 counter jump PSW_RULE comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE") comment \"默认\"" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_SHUNT6 counter jump PSW_RULE comment \"默认\"" [ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(REDIRECT $TCP_REDIR_PORT TPROXY) comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp counter return comment \"默认\"" } echolog " - ${msg2}" } fi # 加载UDP默认代理模式 if [ -n "${UDP_PROXY_MODE}" ]; then [ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ] && { msg2="${msg}使用 UDP 节点[$(config_n_get $UDP_NODE remarks)](TPROXY:${UDP_REDIR_PORT})" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ip daddr $FAKE_IP counter jump PSW_RULE comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_BLACK counter jump PSW_RULE comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_GFW counter jump PSW_RULE comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE") comment \"默认\"" [ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_SHUNT counter jump PSW_RULE comment \"默认\"" [ "${UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(REDIRECT $UDP_REDIR_PORT TPROXY4) comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp counter return comment \"默认\"" [ "$PROXY_IPV6" == "1" ] && { nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"默认\"" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_BLACK6 counter jump PSW_RULE comment \"默认\"" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_GFW6 counter jump PSW_RULE comment \"默认\"" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE") comment \"默认\"" [ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_SHUNT6 counter jump PSW_RULE comment \"默认\"" [ "${UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(REDIRECT $UDP_REDIR_PORT TPROXY) comment \"默认\"" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp counter return comment \"默认\"" } echolog " - ${msg2}" } fi } } filter_haproxy() { for item in ${haproxy_items}; do local ip=$(get_host_ip ipv4 $(echo $item | awk -F ":" '{print $1}') 1) insert_nftset $NFTSET_VPS "-1" $ip done echolog " - [$?]加入负载均衡的节点到nftset[$NFTSET_VPS]直连完成" } filter_vps_addr() { for server_host in $@; do local vps_ip4=$(get_host_ip "ipv4" ${server_host}) local vps_ip6=$(get_host_ip "ipv6" ${server_host}) [ -n "$vps_ip4" ] && insert_nftset $NFTSET_VPS "-1" $vps_ip4 [ -n "$vps_ip6" ] && insert_nftset $NFTSET_VPS6 "-1" $vps_ip6 done } filter_vpsip() { insert_nftset $NFTSET_VPS "-1" $(uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([0-9]{1,3}[\.]){3}[0-9]{1,3}" | grep -v "^127\.0\.0\.1$" | sed -e "/^$/d") echolog " - [$?]加入所有IPv4节点到nftset[$NFTSET_VPS]直连完成" insert_nftset $NFTSET_VPS6 "-1" $(uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "/^$/d") echolog " - [$?]加入所有IPv6节点到nftset[$NFTSET_VPS6]直连完成" } filter_server_port() { local address=${1} local port=${2} local stream=${3} stream=$(echo ${3} | tr 'A-Z' 'a-z') local _is_tproxy _is_tproxy=${is_tproxy} [ "$stream" == "udp" ] && _is_tproxy="TPROXY" for _ipt in 4 6; do [ "$_ipt" == "4" ] && _ip_type=ip [ "$_ipt" == "6" ] && _ip_type=ip6 nft "list chain $NFTABLE_NAME $nft_output_chain" 2>/dev/null | grep -q "${address}:${port}" if [ $? -ne 0 ]; then nft "insert rule $NFTABLE_NAME $nft_output_chain meta l4proto $stream $_ip_type daddr $address $stream dport $port return comment \"${address}:${port}\"" 2>/dev/null fi done } filter_node() { local node=${1} local stream=${2} if [ -n "$node" ]; then local address=$(config_n_get $node address) local port=$(config_n_get $node port) [ -z "$address" ] && [ -z "$port" ] && { return 1 } filter_server_port $address $port $stream filter_server_port $address $port $stream fi } filter_direct_node_list() { [ ! -s "$TMP_PATH/direct_node_list" ] && return for _node_id in $(cat $TMP_PATH/direct_node_list | awk '!seen[$0]++'); do filter_node "$_node_id" TCP filter_node "$_node_id" UDP unset _node_id done } add_firewall_rule() { echolog "开始加载防火墙规则..." gen_nft_tables gen_nftset $NFTSET_VPS ipv4_addr 0 0 gen_nftset $NFTSET_GFW ipv4_addr "2d" 0 gen_nftset $NFTSET_LOCAL ipv4_addr 0 "-1" gen_nftset $NFTSET_LAN ipv4_addr 0 "-1" $(gen_lanlist) if [ -f $RULES_PATH/chnroute.nft ] && [ -s $RULES_PATH/chnroute.nft ] && [ $(awk 'END{print NR}' $RULES_PATH/chnroute.nft) -ge 8 ]; then #echolog "使用缓存加载chnroute..." nft -f $RULES_PATH/chnroute.nft else gen_nftset $NFTSET_CHN ipv4_addr "2d" 0 $(cat $RULES_PATH/chnroute | tr -s '\n' | grep -v "^#") fi gen_nftset $NFTSET_BLACK ipv4_addr "2d" 0 gen_nftset $NFTSET_WHITE ipv4_addr "2d" 0 gen_nftset $NFTSET_BLOCK ipv4_addr "2d" 0 gen_nftset $NFTSET_SHUNT ipv4_addr "2d" 0 gen_nftset $NFTSET_VPS6 ipv6_addr 0 0 gen_nftset $NFTSET_GFW6 ipv6_addr "2d" 0 gen_nftset $NFTSET_LOCAL6 ipv6_addr 0 "-1" gen_nftset $NFTSET_LAN6 ipv6_addr 0 "-1" $(gen_lanlist_6) if [ -f $RULES_PATH/chnroute6.nft ] && [ -s $RULES_PATH/chnroute6.nft ] && [ $(awk 'END{print NR}' $RULES_PATH/chnroute6.nft) -ge 8 ]; then #echolog "使用缓存加载chnroute6..." nft -f $RULES_PATH/chnroute6.nft else gen_nftset $NFTSET_CHN6 ipv6_addr "2d" 0 $(cat $RULES_PATH/chnroute6 | tr -s '\n' | grep -v "^#") fi gen_nftset $NFTSET_BLACK6 ipv6_addr "2d" 0 gen_nftset $NFTSET_WHITE6 ipv6_addr "2d" 0 gen_nftset $NFTSET_BLOCK6 ipv6_addr "2d" 0 gen_nftset $NFTSET_SHUNT6 ipv6_addr "2d" 0 #导入规则列表、分流规则中的IP列表 local USE_SHUNT_NODE=0 local USE_PROXY_LIST_ALL=${USE_PROXY_LIST} local USE_DIRECT_LIST_ALL=${USE_DIRECT_LIST} local USE_BLOCK_LIST_ALL=${USE_BLOCK_LIST} local _TCP_NODE=$(config_t_get global tcp_node) local _UDP_NODE=$(config_t_get global udp_node) local USE_GEOVIEW=$(config_t_get global_rules enable_geoview) [ -n "$_TCP_NODE" ] && [ "$(config_n_get $_TCP_NODE protocol)" = "_shunt" ] && USE_SHUNT_TCP=1 && USE_SHUNT_NODE=1 [ -n "$_UDP_NODE" ] && [ "$(config_n_get $_UDP_NODE protocol)" = "_shunt" ] && USE_SHUNT_UDP=1 && USE_SHUNT_NODE=1 [ "$_UDP_NODE" = "tcp" ] && USE_SHUNT_UDP=$USE_SHUNT_TCP for acl_section in $(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do [ "$(config_n_get $acl_section enabled)" != "1" ] && continue [ "$(config_n_get $acl_section use_global_config 0)" != "1" ] && { [ "$(config_n_get $acl_section use_direct_list 1)" = "1" ] && USE_PROXY_LIST_ALL=1 [ "$(config_n_get $acl_section use_proxy_list 1)" = "1" ] && USE_DIRECT_LIST_ALL=1 [ "$(config_n_get $acl_section use_block_list 1)" = "1" ] && USE_BLOCK_LIST_ALL=1 } for _node in $(config_n_get $acl_section tcp_node) $(config_n_get $acl_section udp_node); do local node_protocol=$(config_n_get $_node protocol) [ "$node_protocol" = "_shunt" ] && { USE_SHUNT_NODE=1; break; } done done #直连列表 [ "$USE_DIRECT_LIST_ALL" = "1" ] && { insert_nftset $NFTSET_WHITE "0" $(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_WHITE6 "0" $(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") [ "$USE_GEOVIEW" = "1" ] && { local GEOIP_CODE=$(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then insert_nftset $NFTSET_WHITE "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_WHITE6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") echolog " - [$?]解析并加入[直连列表] GeoIP 到 NFTSET 完成" fi } } #代理列表 [ "$USE_PROXY_LIST_ALL" = "1" ] && { insert_nftset $NFTSET_BLACK "0" $(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_BLACK6 "0" $(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") [ "$USE_GEOVIEW" = "1" ] && { local GEOIP_CODE=$(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then insert_nftset $NFTSET_BLACK "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_BLACK6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") echolog " - [$?]解析并加入[代理列表] GeoIP 到 NFTSET 完成" fi } } #屏蔽列表 [ "$USE_BLOCK_LIST_ALL" = "1" ] && { insert_nftset $NFTSET_BLOCK "0" $(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_BLOCK6 "0" $(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") [ "$USE_GEOVIEW" = "1" ] && { local GEOIP_CODE=$(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then insert_nftset $NFTSET_BLOCK "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_BLOCK6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") echolog " - [$?]解析并加入[屏蔽列表] GeoIP 到 NFTSET 完成" fi } } #分流列表 [ "$USE_SHUNT_NODE" = "1" ] && { local GEOIP_CODE="" local shunt_ids=$(uci show $CONFIG | grep "=shunt_rules" | awk -F '.' '{print $2}' | awk -F '=' '{print $1}') for shunt_id in $shunt_ids; do insert_nftset $NFTSET_SHUNT "0" $(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_SHUNT6 "0" $(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") [ "$USE_GEOVIEW" = "1" ] && { local geoip_code=$(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') [ -n "$geoip_code" ] && GEOIP_CODE="${GEOIP_CODE:+$GEOIP_CODE,}$geoip_code" } done if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then insert_nftset $NFTSET_SHUNT "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}") insert_nftset $NFTSET_SHUNT6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") echolog " - [$?]解析并加入[分流节点] GeoIP 到 NFTSET 完成" fi } insert_nftset $NFTSET_LOCAL "-1" $(ip address show | grep -w "inet" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/ /\n/g") insert_nftset $NFTSET_LOCAL6 "-1" $(ip address show | grep -w "inet6" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/ /\n/g") # 忽略特殊IP段 local lan_ifname lan_ip lan_ifname=$(uci -q -p /tmp/state get network.lan.ifname) [ -n "$lan_ifname" ] && { lan_ip=$(ip address show $lan_ifname | grep -w "inet" | awk '{print $2}') lan_ip6=$(ip address show $lan_ifname | grep -w "inet6" | awk '{print $2}') #echolog "本机IPv4网段互访直连:${lan_ip}" #echolog "本机IPv6网段互访直连:${lan_ip6}" [ -n "$lan_ip" ] && insert_nftset $NFTSET_LAN "-1" $(echo $lan_ip | sed -e "s/ /\n/g") [ -n "$lan_ip6" ] && insert_nftset $NFTSET_LAN6 "-1" $(echo $lan_ip6 | sed -e "s/ /\n/g") } [ -n "$ISP_DNS" ] && { #echolog "处理 ISP DNS 例外..." for ispip in $ISP_DNS; do insert_nftset $NFTSET_WHITE 0 $ispip echolog " - [$?]追加ISP IPv4 DNS到白名单:${ispip}" done } [ -n "$ISP_DNS6" ] && { #echolog "处理 ISP IPv6 DNS 例外..." for ispip6 in $ISP_DNS6; do insert_nftset $NFTSET_WHITE6 0 $ispip6 echolog " - [$?]追加ISP IPv6 DNS到白名单:${ispip6}" done } # 过滤所有节点IP filter_vpsip > /dev/null 2>&1 & # filter_haproxy > /dev/null 2>&1 & # Prevent some conditions filter_vps_addr $(config_n_get $TCP_NODE address) $(config_n_get $UDP_NODE address) > /dev/null 2>&1 & filter_vps_addr $(config_n_get $TCP_NODE download_address) $(config_n_get $UDP_NODE download_address) > /dev/null 2>&1 & accept_icmp=$(config_t_get global_forwarding accept_icmp 0) accept_icmpv6=$(config_t_get global_forwarding accept_icmpv6 0) if [ "${TCP_PROXY_WAY}" = "redirect" ]; then unset is_tproxy nft_prerouting_chain="PSW_NAT" nft_output_chain="PSW_OUTPUT_NAT" elif [ "${TCP_PROXY_WAY}" = "tproxy" ]; then is_tproxy="TPROXY" nft_prerouting_chain="PSW_MANGLE" nft_output_chain="PSW_OUTPUT_MANGLE" fi nft "add chain $NFTABLE_NAME PSW_DIVERT" nft "flush chain $NFTABLE_NAME PSW_DIVERT" nft "add rule $NFTABLE_NAME PSW_DIVERT meta l4proto tcp socket transparent 1 mark set 1 counter accept" nft "add chain $NFTABLE_NAME PSW_DNS" nft "flush chain $NFTABLE_NAME PSW_DNS" if [ $(config_t_get global dns_redirect "1") = "0" ]; then #Only hijack when dest address is local IP nft "insert rule $NFTABLE_NAME dstnat ip daddr @${NFTSET_LOCAL} jump PSW_DNS" nft "insert rule $NFTABLE_NAME dstnat ip6 daddr @${NFTSET_LOCAL6} jump PSW_DNS" else nft "insert rule $NFTABLE_NAME dstnat jump PSW_DNS" fi # for ipv4 ipv6 tproxy mark nft "add chain $NFTABLE_NAME PSW_RULE" nft "flush chain $NFTABLE_NAME PSW_RULE" nft "add rule $NFTABLE_NAME PSW_RULE meta mark set ct mark counter" nft "add rule $NFTABLE_NAME PSW_RULE meta mark 1 counter return" nft "add rule $NFTABLE_NAME PSW_RULE tcp flags &(fin|syn|rst|ack) == syn meta mark set mark and 0x0 xor 0x1 counter" nft "add rule $NFTABLE_NAME PSW_RULE meta l4proto udp ct state new meta mark set mark and 0x0 xor 0x1 counter" nft "add rule $NFTABLE_NAME PSW_RULE ct mark set mark counter" #ipv4 tproxy mode and udp nft "add chain $NFTABLE_NAME PSW_MANGLE" nft "flush chain $NFTABLE_NAME PSW_MANGLE" nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_LAN counter return" nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_VPS counter return" nft "add chain $NFTABLE_NAME PSW_OUTPUT_MANGLE" nft "flush chain $NFTABLE_NAME PSW_OUTPUT_MANGLE" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_LAN counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_VPS counter return" [ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_BLOCK counter drop" [ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_WHITE counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE meta mark 0xff counter return" # jump chains nft "add rule $NFTABLE_NAME mangle_prerouting ip protocol udp counter jump PSW_MANGLE" [ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME mangle_prerouting ip protocol tcp counter jump PSW_MANGLE" insert_rule_before "$NFTABLE_NAME" "mangle_prerouting" "PSW_MANGLE" "counter jump PSW_DIVERT" #ipv4 tcp redirect mode [ -z "${is_tproxy}" ] && { nft "add chain $NFTABLE_NAME PSW_NAT" nft "flush chain $NFTABLE_NAME PSW_NAT" nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_LAN counter return" nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_VPS counter return" nft "add rule $NFTABLE_NAME dstnat ip protocol tcp counter jump PSW_NAT" nft "add chain $NFTABLE_NAME PSW_OUTPUT_NAT" nft "flush chain $NFTABLE_NAME PSW_OUTPUT_NAT" nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_LAN counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_VPS counter return" [ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_BLOCK counter drop" [ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_WHITE counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT meta mark 0xff counter return" } #icmp ipv6-icmp redirect if [ "$accept_icmp" = "1" ]; then nft "add chain $NFTABLE_NAME PSW_ICMP_REDIRECT" nft "flush chain $NFTABLE_NAME PSW_ICMP_REDIRECT" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip daddr @$NFTSET_LAN counter return" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip daddr @$NFTSET_VPS counter return" [ "$accept_icmpv6" = "1" ] && { nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip6 daddr @$NFTSET_LAN6 counter return" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip6 daddr @$NFTSET_VPS6 counter return" } nft "add rule $NFTABLE_NAME dstnat meta l4proto {icmp,icmpv6} counter jump PSW_ICMP_REDIRECT" nft "add rule $NFTABLE_NAME nat_output meta l4proto {icmp,icmpv6} counter jump PSW_ICMP_REDIRECT" fi WAN_IP=$(get_wan_ip) if [ -n "${WAN_IP}" ]; then [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ip daddr ${WAN_IP} counter return comment \"WAN_IP_RETURN\"" nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr ${WAN_IP} counter return comment \"WAN_IP_RETURN\"" echolog " - [$?]追加WAN IP到nftables:${WAN_IP}" fi unset WAN_IP ip rule add fwmark 1 lookup 100 ip route add local 0.0.0.0/0 dev lo table 100 #ipv6 tproxy mode and udp nft "add chain $NFTABLE_NAME PSW_MANGLE_V6" nft "flush chain $NFTABLE_NAME PSW_MANGLE_V6" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_LAN6 counter return" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_VPS6 counter return" nft "add chain $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6" nft "flush chain $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_LAN6 counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_VPS6 counter return" [ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_BLOCK6 counter drop" [ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_WHITE6 counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta mark 0xff counter return" [ -n "$IPT_APPEND_DNS" ] && { local local_dns dns_address dns_port for local_dns in $(echo $IPT_APPEND_DNS | tr ',' ' '); do dns_address=$(echo "$local_dns" | sed -E 's/(@|\[)?([0-9a-fA-F:.]+)(@|#|$).*/\2/') dns_port=$(echo "$local_dns" | sed -nE 's/.*#([0-9]+)$/\1/p') if echo "$dns_address" | grep -q -v ':'; then nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr ${dns_address} $(factor ${dns_port:-53} "udp dport") counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol tcp ip daddr ${dns_address} $(factor ${dns_port:-53} "tcp dport") counter return" echolog " - [$?]追加直连DNS到nftables:${dns_address}:${dns_port:-53}" else nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr ${dns_address} $(factor ${dns_port:-53} "udp dport") counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr ${dns_address} $(factor ${dns_port:-53} "tcp dport") counter return" echolog " - [$?]追加直连DNS到nftables:[${dns_address}]:${dns_port:-53}" fi done } # jump chains [ "$PROXY_IPV6" == "1" ] && { nft "add rule $NFTABLE_NAME mangle_prerouting meta nfproto {ipv6} counter jump PSW_MANGLE_V6" nft "add rule $NFTABLE_NAME mangle_output meta nfproto {ipv6} counter jump PSW_OUTPUT_MANGLE_V6 comment \"PSW_OUTPUT_MANGLE\"" WAN6_IP=$(get_wan6_ip) [ -n "${WAN6_IP}" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr ${WAN6_IP} counter return comment \"WAN6_IP_RETURN\"" unset WAN6_IP ip -6 rule add fwmark 1 table 100 ip -6 route add local ::/0 dev lo table 100 } [ "$TCP_UDP" = "1" ] && [ -z "$UDP_NODE" ] && UDP_NODE=$TCP_NODE [ "$ENABLED_DEFAULT_ACL" == 1 ] && { msg="【路由器本机】," [ "$TCP_NO_REDIR_PORTS" != "disable" ] && { nft "add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return" if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]" else unset LOCALHOST_TCP_PROXY_MODE echolog " - ${msg}不代理所有 TCP 端口" fi } [ "$UDP_NO_REDIR_PORTS" != "disable" ] && { nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return" nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return" if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]" else unset LOCALHOST_UDP_PROXY_MODE echolog " - ${msg}不代理所有 UDP 端口" fi } if ([ -n "$TCP_NODE" ] && [ -n "${LOCALHOST_TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${LOCALHOST_UDP_PROXY_MODE}" ]); then [ -n "$DNS_REDIRECT_PORT" ] && { nft "add rule $NFTABLE_NAME nat_output ip protocol udp oif lo udp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\"" nft "add rule $NFTABLE_NAME nat_output ip protocol tcp oif lo tcp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\"" nft "add rule $NFTABLE_NAME nat_output meta l4proto udp oif lo udp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\"" nft "add rule $NFTABLE_NAME nat_output meta l4proto tcp oif lo tcp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\"" } fi [ -n "${LOCALHOST_TCP_PROXY_MODE}" -o -n "${LOCALHOST_UDP_PROXY_MODE}" ] && { [ "$TCP_PROXY_DROP_PORTS" != "disable" ] && { nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr $FAKE_IP $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop [ "${USE_PROXY_LIST}" = "1" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_BLACK $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop [ "${USE_GFW_LIST}" = "1" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_GFW $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop [ "${CHN_LIST}" != "0" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_CHN $(factor $TCP_PROXY_DROP_PORTS "tcp dport") $(get_jump_ipt ${CHN_LIST} "counter drop") [ "${USE_SHUNT_TCP}" = "1" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_SHUNT $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]" } [ "$UDP_PROXY_DROP_PORTS" != "disable" ] && { nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr $FAKE_IP $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop [ "${USE_PROXY_LIST}" = "1" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_BLACK $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop [ "${USE_GFW_LIST}" = "1" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_GFW $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop [ "${CHN_LIST}" != "0" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_CHN $(factor $UDP_PROXY_DROP_PORTS "udp dport") $(get_jump_ipt ${CHN_LIST} "counter drop") [ "${USE_SHUNT_UDP}" = "1" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_SHUNT $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop [ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE counter ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]" } } # 加载路由器自身代理 TCP if [ -n "$TCP_NODE" ]; then _proxy_tcp_access() { [ -n "${2}" ] || return 0 if echo "${2}" | grep -q -v ':'; then nft "get element $NFTABLE_NAME $NFTSET_LAN {${2}}" &>/dev/null [ $? -eq 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问" return 0 } if [ -z "${is_tproxy}" ]; then nft insert rule $NFTABLE_NAME PSW_OUTPUT_NAT ip protocol tcp ip daddr ${2} tcp dport ${3} $(REDIRECT $TCP_REDIR_PORT) else nft insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol tcp ip daddr ${2} tcp dport ${3} counter jump PSW_RULE nft insert rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp iif lo tcp dport ${3} ip daddr ${2} $(REDIRECT $TCP_REDIR_PORT TPROXY4) comment \"本机\" fi echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 TCP 转发链" else nft "get element $NFTABLE_NAME $NFTSET_LAN6 {${2}}" &>/dev/null [ $? -eq 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问" return 0 } nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr ${2} tcp dport ${3} counter jump PSW_RULE" nft "insert rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo tcp dport ${3} ip6 daddr ${2} $(REDIRECT $TCP_REDIR_PORT TPROXY6) comment \"本机\"" echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 TCP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!" fi } [ "$TCP_PROXY_DNS" == 1 ] && hosts_foreach REMOTE_DNS _proxy_tcp_access 53 [ "$accept_icmp" = "1" ] && { nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr $FAKE_IP counter redirect" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_BLACK counter redirect" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_GFW counter redirect" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST})" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_SHUNT counter redirect" [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp counter redirect" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp counter return" } [ "$accept_icmpv6" = "1" ] && { nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr $FAKE_IP_6 counter redirect" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_BLACK6 counter redirect" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_GFW6 counter redirect" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST})" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_SHUNT6 counter redirect" [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 counter redirect" nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 counter return" } if [ -n "${is_tproxy}" ]; then nft_chain="PSW_OUTPUT_MANGLE" nft_j="counter jump PSW_RULE" else nft_chain="PSW_OUTPUT_NAT" nft_j="$(REDIRECT $TCP_REDIR_PORT)" fi [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && { nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr $FAKE_IP ${nft_j}" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_BLACK $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_GFW $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_CHN $(factor $TCP_REDIR_PORTS "tcp dport") $(get_jump_ipt ${CHN_LIST} "${nft_j}")" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_SHUNT $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}" [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}" [ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp iif lo $(REDIRECT $TCP_REDIR_PORT TPROXY4) comment \"本机\"" } [ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp iif lo counter return comment \"本机\"" [ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME mangle_output ip protocol tcp counter jump PSW_OUTPUT_MANGLE comment \"PSW_OUTPUT_MANGLE\"" [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME nat_output ip protocol tcp counter jump PSW_OUTPUT_NAT" [ "$PROXY_IPV6" == "1" ] && { [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && { nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_BLACK6 $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_GFW6 $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_CHN6 $(factor $TCP_REDIR_PORTS "tcp dport") $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE")" [ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_SHUNT6 $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE" [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo $(REDIRECT $TCP_REDIR_PORT TPROXY) comment \"本机\"" } nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo counter return comment \"本机\"" } fi # 加载路由器自身代理 UDP if [ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ]; then _proxy_udp_access() { [ -n "${2}" ] || return 0 if echo "${2}" | grep -q -v ':'; then nft "get element $NFTABLE_NAME $NFTSET_LAN {${2}}" &>/dev/null [ $? == 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问" return 0 } nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr ${2} udp dport ${3} counter jump PSW_RULE" nft "insert rule $NFTABLE_NAME PSW_MANGLE ip protocol udp iif lo ip daddr ${2} $(REDIRECT $UDP_REDIR_PORT TPROXY4) comment \"本机\"" echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 UDP 转发链" else nft "get element $NFTABLE_NAME $NFTSET_LAN6 {${2}}" &>/dev/null [ $? == 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问" return 0 } nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr ${2} udp dport ${3} counter jump PSW_RULE" nft "insert rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo ip6 daddr ${2} $(REDIRECT $UDP_REDIR_PORT TPROXY6) comment \"本机\"" echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 UDP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!" fi } [ -n "${UDP_PROXY_DNS}" ] && hosts_foreach REMOTE_DNS _proxy_udp_access 53 [ -n "${LOCALHOST_UDP_PROXY_MODE}" ] && { nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr $FAKE_IP counter jump PSW_RULE" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_BLACK $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_GFW $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_CHN $(factor $UDP_REDIR_PORTS "udp dport") $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE")" [ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_SHUNT $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" [ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp iif lo $(REDIRECT $UDP_REDIR_PORT TPROXY4) comment \"本机\"" } nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp iif lo counter return comment \"本机\"" nft "add rule $NFTABLE_NAME mangle_output ip protocol udp counter jump PSW_OUTPUT_MANGLE comment \"PSW_OUTPUT_MANGLE\"" [ "$PROXY_IPV6" == "1" ] && { [ -n "${LOCALHOST_UDP_PROXY_MODE}" ] && { nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE" [ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_BLACK6 $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_GFW6 $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" [ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_CHN6 $(factor $UDP_REDIR_PORTS "udp dport") $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE")" [ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_SHUNT6 $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" [ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE" nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp iif lo $(REDIRECT $UDP_REDIR_PORT TPROXY) comment \"本机\"" } nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp iif lo counter return comment \"本机\"" } fi nft "add rule $NFTABLE_NAME mangle_output oif lo counter return comment \"PSW_OUTPUT_MANGLE\"" nft "add rule $NFTABLE_NAME mangle_output meta mark 1 counter return comment \"PSW_OUTPUT_MANGLE\"" } # 加载ACLS load_acl [ -d "${TMP_IFACE_PATH}" ] && { for iface in $(ls ${TMP_IFACE_PATH}); do nft "insert rule $NFTABLE_NAME $nft_output_chain oif $iface counter return" nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 oif $iface counter return" done } filter_direct_node_list > /dev/null 2>&1 & echolog "防火墙规则加载完成!" } del_firewall_rule() { for nft in "dstnat" "srcnat" "nat_output" "mangle_prerouting" "mangle_output"; do local handles=$(nft -a list chain $NFTABLE_NAME ${nft} 2>/dev/null | grep -E "PSW_" | awk -F '# handle ' '{print$2}') for handle in $handles; do nft delete rule $NFTABLE_NAME ${nft} handle ${handle} 2>/dev/null done done for handle in $(nft -a list chains | grep -E "chain PSW_" | grep -v "PSW_RULE" | awk -F '# handle ' '{print$2}'); do nft delete chain $NFTABLE_NAME handle ${handle} 2>/dev/null done # Need to be removed at the end, otherwise it will show "Resource busy" nft delete chain $NFTABLE_NAME handle $(nft -a list chains | grep -E "PSW_RULE" | awk -F '# handle ' '{print$2}') 2>/dev/null ip rule del fwmark 1 lookup 100 2>/dev/null ip route del local 0.0.0.0/0 dev lo table 100 2>/dev/null ip -6 rule del fwmark 1 table 100 2>/dev/null ip -6 route del local ::/0 dev lo table 100 2>/dev/null destroy_nftset $NFTSET_LOCAL destroy_nftset $NFTSET_LAN destroy_nftset $NFTSET_VPS destroy_nftset $NFTSET_SHUNT #destroy_nftset $NFTSET_GFW #destroy_nftset $NFTSET_CHN destroy_nftset $NFTSET_BLACK destroy_nftset $NFTSET_BLOCK destroy_nftset $NFTSET_WHITE destroy_nftset $NFTSET_LOCAL6 destroy_nftset $NFTSET_LAN6 destroy_nftset $NFTSET_VPS6 destroy_nftset $NFTSET_SHUNT6 #destroy_nftset $NFTSET_GFW6 #destroy_nftset $NFTSET_CHN6 destroy_nftset $NFTSET_BLACK6 destroy_nftset $NFTSET_BLOCK6 destroy_nftset $NFTSET_WHITE6 $DIR/app.sh echolog "删除 nftables 规则完成。" } flush_nftset() { $DIR/app.sh echolog "清空 NFTSet。" for _name in $(nft -a list sets | grep -E "passwall" | awk -F 'set ' '{print $2}' | awk '{print $1}'); do destroy_nftset ${_name} done } flush_table() { nft flush table $NFTABLE_NAME nft delete table $NFTABLE_NAME } flush_include() { echo '#!/bin/sh' >$FWI } gen_include() { flush_include local nft_chain_file=$TMP_PATH/PSW_RULE.nft echo '#!/usr/sbin/nft -f' > $nft_chain_file nft list table $NFTABLE_NAME >> $nft_chain_file local __nft=" " __nft=$(cat <<- EOF [ -z "\$(nft list chain $NFTABLE_NAME mangle_prerouting | grep PSW_DIVERT)" ] && nft -f ${nft_chain_file} [ -z "${is_tproxy}" ] && { PR_INDEX=\$(sh ${MY_PATH} RULE_LAST_INDEX "$NFTABLE_NAME" PSW_NAT WAN_IP_RETURN -1) if [ \$PR_INDEX -ge 0 ]; then WAN_IP=\$(sh ${MY_PATH} get_wan_ip) [ ! -z "\${WAN_IP}" ] && nft "replace rule $NFTABLE_NAME PSW_NAT handle \$PR_INDEX ip daddr "\${WAN_IP}" counter return comment \"WAN_IP_RETURN\"" fi } PR_INDEX=\$(sh ${MY_PATH} RULE_LAST_INDEX "$NFTABLE_NAME" PSW_MANGLE WAN_IP_RETURN -1) if [ \$PR_INDEX -ge 0 ]; then WAN_IP=\$(sh ${MY_PATH} get_wan_ip) [ ! -z "\${WAN_IP}" ] && nft "replace rule $NFTABLE_NAME PSW_MANGLE handle \$PR_INDEX ip daddr "\${WAN_IP}" counter return comment \"WAN_IP_RETURN\"" fi [ "$PROXY_IPV6" == "1" ] && { PR_INDEX=\$(sh ${MY_PATH} RULE_LAST_INDEX "$NFTABLE_NAME" PSW_MANGLE_V6 WAN6_IP_RETURN -1) if [ \$PR_INDEX -ge 0 ]; then WAN6_IP=\$(sh ${MY_PATH} get_wan6_ip) [ ! -z "\${WAN_IP}" ] && nft "replace rule $NFTABLE_NAME PSW_MANGLE_V6 handle \$PR_INDEX ip6 daddr "\${WAN6_IP}" counter return comment \"WAN6_IP_RETURN\"" fi } EOF ) cat <<-EOF >> $FWI ${__nft} return 0 EOF return 0 } start() { [ "$ENABLED_DEFAULT_ACL" == 0 -a "$ENABLED_ACLS" == 0 ] && return add_firewall_rule gen_include } stop() { del_firewall_rule [ $(config_t_get global flush_set "0") = "1" ] && { uci -q delete ${CONFIG}.@global[0].flush_set uci -q commit ${CONFIG} #flush_table flush_nftset rm -rf /tmp/etc/passwall_tmp/singbox* rm -rf /tmp/etc/passwall_tmp/dnsmasq* } flush_include } arg1=$1 shift case $arg1 in RULE_LAST_INDEX) RULE_LAST_INDEX "$@" ;; insert_rule_before) insert_rule_before "$@" ;; insert_rule_after) insert_rule_after "$@" ;; get_wan_ip) get_wan_ip ;; get_wan6_ip) get_wan6_ip ;; filter_direct_node_list) filter_direct_node_list ;; stop) stop ;; start) start ;; *) ;; esac
294coder/Efficient-MIF
3,551
Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_ref_unbatched_images.m
function res = analysis_ref_unbatched_images(dir_path, ratio, sensor, flag_cut_bounds, printout) % data should be [0, max_range] % e.g. wv3: max_range: 2047 if nargin == 3 flag_cut_bounds = 0; printout = 1; elseif nargin == 4 printout = 1; end sensor = lower(sensor); if sensor == "wv3" || sensor == "wv2" || sensor == "qb" const = 2047; gt_key = 'gt'; if sensor == "wv3" C = 8; gt_path = "/Data2/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5"; elseif sensor == "wv2" C = 8; gt_path = '/Data2/ZiHanCao/datasets/pansharpening/wv2/reduced_examples/test_wv2_multiExm1.h5'; else C = 4; gt_path = "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"; end elseif sensor == "gf2" || sensor == "gf" const = 1023; gt_key = 'gt'; C = 4; gt_path = "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5"; elseif sensor == "cave" || sensor == "harvard" const = 1; gt_key = 'GT'; C = 31; if sensor == "cave" if ratio == 4 gt_path = "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5"; elseif ratio == 8 gt_path = "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/test_cave(with_up)x8_rgb.h5"; end else % % harvard gt_path = ''; end else error(strcat(sensor, ' is not supported!')) end gts = h5read(gt_path, strcat('/', gt_key)); gts = permute(gts, [4, 2, 1, 3]); bs = size(gts, 1); addpath('./Tools') addpath('./Quality_Indices/') Q_block_size = 32; thvalues = 0; L = 11; full_res = 0; % default to 0 if full_res dim_cut = 21; else dim_cut = 30; end res = {}; sam = zeros(1, bs); ergas = zeros(1, bs); scc = zeros(1, bs); qn = zeros(1, bs); qave = zeros(1, bs); psnr = zeros(1, bs); ssim = zeros(1, bs); for i = (0:bs - 1) p = strcat(dir_path, "/", "output_mulExm_", string(i), ".mat"); data = load(p); sr_key = fieldnames(data); sr_key = sr_key{1}; sr1 = data.(string(sr_key)); % check the size if size(sr1, 1) == 1 && length(size(sr1)) == 4 sr1 = squeeze(sr1); end if size(sr1, 3) ~= C && length(size(sr1)) == 3 sr1 = permute(sr1, [2, 3, 1]); end i = i + 1; gt1 = squeeze(gts(i, :, :, :)); [q_avg_gt, sam_gt, ergas_gt, scc_gt, q_gt] = indexes_evaluation(sr1, gt1, ratio, L, Q_block_size, flag_cut_bounds, dim_cut, thvalues); [psnr_gt, ssim_gt] = quality_assess(sr1 / const, gt1 / const); sam(i) = sam_gt; ergas(i) = ergas_gt; scc(i) = scc_gt; qn(i) = q_gt; qave(i) = q_avg_gt; psnr(i) = psnr_gt; ssim(i) = ssim_gt; if printout fprintf("sample %d - sam: %f, ergas: %f, scc: %f, qn: %f, q_ave: %f, psnr: %f, ssim: %f \n", i, sam_gt, ergas_gt, scc_gt, q_gt, q_avg_gt, psnr_gt, ssim_gt) end end res.sam = [mean(sam), std(sam)]; res.ergas = [mean(ergas), std(ergas)]; res.scc = [mean(scc), std(scc)]; res.qn = [mean(qn), std(qn)]; res.qave = [mean(qave), std(qave)]; res.psnr = [mean(psnr), std(psnr)]; res.ssim = [mean(ssim), std(ssim)]; end
294coder/Efficient-MIF
4,692
task_datasets/GF2.py
import torch import torch.utils.data as data import torchvision.transforms as T import cv2 import numpy as np import h5py from typing import List, Tuple, Optional class Identity: def __call__(self, *args): return args class GF2Datasets(data.Dataset): def __init__( self, d, aug_prob=0.0, hp=False, hp_ksize=(5, 5), norm_range=True, full_res=False, const=1023.0, ): """ :param d: h5py.File or dict or path :param aug_prob: augmentation probability :param hp: high pass for ms and pan. x = x - cv2.boxFilter(x) :param hp_ksize: cv2.boxFiler kernel size :param norm_range: normalize data range """ super(GF2Datasets, self).__init__() # FIXME: should pass @path rather than @file which is h5py.File object to avoid can not be pickled error if isinstance(d, (str, h5py.File)): if isinstance(d, str): d = h5py.File(d) print( "warning: when @file is a h5py.File object, it can not be pickled.", "try to set DataLoader number_worker to 0", ) if not full_res: self.gt, self.ms, self.lms, self.pan = self.get_divided(d) print("datasets shape:") print("{:^20}{:^20}{:^20}{:^20}".format("pan", "ms", "lms", "gt")) print( "{:^20}{:^20}{:^20}{:^20}".format( str(self.pan.shape), str(self.ms.shape), str(self.lms.shape), str(self.gt.shape), ) ) else: self.ms, self.lms, self.pan = self.get_divided(d, True) print("datasets shape:") print("{:^20}{:^20}{:^20}".format("pan", "ms", "lms")) print( "{:^20}{:^20}{:^20}".format( str(self.pan.shape), str(self.ms.shape), str(self.lms.shape) ) ) self.size = self.ms.shape[0] # highpass filter self.hp = hp self.hp_ksize = hp_ksize if hp and hp_ksize is not None: self.group_high_pass(hp_ksize) # to tensor if norm_range: def norm_func(x): # return torch.tensor(x) / 2047. return torch.tensor(x, dtype=torch.float32) / const else: def norm_func(x): return torch.tensor(x, dtype=torch.float32) self.pan = norm_func(self.pan) self.ms = norm_func(self.ms) self.lms = norm_func(self.lms) if not full_res: self.gt = norm_func(self.gt) # geometrical transformation self.aug_prob = aug_prob self.geo_trans = ( T.Compose( [T.RandomVerticalFlip(p=aug_prob), T.RandomHorizontalFlip(p=aug_prob)] ) if aug_prob != 0.0 else Identity() ) @staticmethod def get_divided(d, full_resolution=False): if not full_resolution: return ( np.asarray(d["gt"]), np.asarray(d["ms"]), np.asarray(d["lms"]), np.asarray(d["pan"]), ) else: return (np.asarray(d["ms"]), np.asarray(d["lms"]), np.asarray(d["pan"])) @staticmethod def _get_high_pass(data, k_size): for i, img in enumerate(data): hp = cv2.boxFilter(img.transpose(1, 2, 0), -1, k_size) if hp.ndim == 2: hp = hp[..., np.newaxis] data[i] = img - hp.transpose(2, 0, 1) return data def group_high_pass(self, k_size): self.ms = self._get_high_pass(self.ms, k_size) self.pan = self._get_high_pass(self.pan, k_size) def aug_trans(self, *data): data_list = [] seed = torch.random.seed() for d in data: torch.manual_seed(seed) d = self.geo_trans(d) data_list.append(d) return data_list def __getitem__(self, item): if hasattr(self, "gt"): tuple_data = (self.pan[item], self.ms[item], self.lms[item], self.gt[item]) else: tuple_data = (self.pan[item], self.ms[item], self.lms[item]) return self.aug_trans(*tuple_data) if self.aug_prob != 0.0 else tuple_data def __len__(self): return self.size def __repr__(self): return ( f"num: {self.size} \n " f"augmentation: {self.geo_trans} \n" f"get high pass ms and pan: {self.hp} \n " f"filter kernel size: {self.hp_ksize}" )
294coder/Efficient-MIF
6,408
task_datasets/HISR.py
import time from typing import Union import matplotlib.pyplot as plt import torch import torch.utils.data as data import torchvision.transforms as T import h5py import random from typing import List, Tuple, Optional, Callable def default_dataset_fn(*x): return x[0] class HISRDatasets(data.Dataset): # FIXME: when use this Dataset, you should set num_works to 0 or it will raise unpickable error def __init__( self, file: Union[h5py.File, str, dict], aug_prob=0.0, rgb_to_bgr=False, full_res=False, *, dataset_fn=None ): super(HISRDatasets, self).__init__() # warning: you should not save file (h5py.File) in this class, # or it will raise CAN NOT BE PICKLED error in multiprocessing # FIXME: should pass @path rather than @file which is h5py.File object to avoid can not be pickled error if isinstance(file, (str, h5py.File)): if isinstance(file, str): file = h5py.File(file) print( "warning: when @file is a h5py.File object, it can not be pickled.", "try to set DataLoader number_worker to 0", ) # checking dataset_fn type if dataset_fn is not None: if isinstance(dataset_fn, (list, tuple)): def _apply_fn(tensor): for fn in dataset_fn: tensor = fn(tensor) return tensor self.dataset_fn = _apply_fn elif isinstance(dataset_fn, Callable): self.dataset_fn = dataset_fn else: raise TypeError("dataset_fn should be a list of callable or a callable object") else: self.dataset_fn = default_dataset_fn self.full_res = full_res data_s= self._split_parts( file, rgb_to_bgr=rgb_to_bgr, full=full_res ) if len(data_s) == 4: self.gt, self.lr_hsi, self.rgb, self.hsi_up = data_s else: self.lr_hsi, self.rgb, self.hsi_up = data_s self.size = self.rgb.shape[-2:] print("dataset shape:") # print dataset info if not full_res: print("{:^20}{:^20}{:^20}{:^20}".format("lr_hsi", "hsi_up", "rgb", "gt")) print( "{:^20}{:^20}{:^20}{:^20}".format( str(tuple(self.lr_hsi.shape)), str(tuple(self.hsi_up.shape)), str(tuple(self.rgb.shape)), str(tuple(self.gt.shape)), ) ) else: print("{:^20}{:^20}{:^20}".format("lr_hsi", "hsi_up", "rgb")) print( "{:^20}{:^20}{:^20}".format( str(tuple(self.lr_hsi.shape)), str(tuple(self.hsi_up.shape)), str(tuple(self.rgb.shape)), ) ) # geometrical transformation self.aug_prob = aug_prob self.geo_trans = ( T.Compose( [ # T.RandomHorizontalFlip(p=self.aug_prob), # T.RandomVerticalFlip(p=self.aug_prob), T.RandomApply( [ T.RandomErasing( p=self.aug_prob, scale=(0.02, 0.15), ratio=(0.2, 1.0) ), T.RandomAffine( degrees=(0, 70), translate=(0.1, 0.2), scale=(0.95, 1.2), interpolation=T.InterpolationMode.BILINEAR, ), ], p=self.aug_prob, ), # T.RandomAutocontrast(p=self.aug_prob), # T.RandomAdjustSharpness(sharpness_factor=2, p=self.aug_prob) # T.RandomErasing(p=self.aug_prob) ] ) if aug_prob != 0.0 else lambda *x: x ) def _split_parts(self, file, load_all=True, rgb_to_bgr=False, keys=None, full=False): # has already been normalized # warning: key RGB is HRMSI when the dataset is GF5-GF1 if not full: keys = ['GT', 'LRHSI', 'RGB', 'HSI_up'] else: keys = ['LRHSI', 'RGB', 'HSI_up'] if load_all: # load all data in memory data = [] for k in keys: data.append( self.dataset_fn(torch.tensor(file[k][:], dtype=torch.float32)), ) if rgb_to_bgr: print("warning: rgb to bgr, for testing generalization only.") # rgb -> bgr if not full: data[2] = data[2][:, [-1, 1, 0]] else: data[1] = data[1][:, [-1, 1, 0]] return data else: # warning: it will ignore @normalize # warning: "GT" can not be access in FULL mode return ( file.get("GT"), file.get("LRHSI"), file.get("RGB"), file.get("HSI_up"), ) def aug_trans(self, *data): data_list = [] seed = torch.random.seed() for d in data: torch.manual_seed(seed) random.seed(seed) d = self.geo_trans(d) data_list.append(d) return tuple(data_list) def __getitem__(self, index): # gt: [31, 64, 64] # lr_hsi: [31, 16, 16] # rbg: [3, 64, 64] # hsi_up: [31, 64, 64] # harvard [rgb] # cave [bgr] if not self.full_res: tuple_data = ( self.rgb[index], self.lr_hsi[index], self.hsi_up[index], self.gt[index], ) else: tuple_data = ( self.rgb[index], self.lr_hsi[index], self.hsi_up[index], ) if self.aug_prob != 0.0: return self.aug_trans(*tuple_data) else: return tuple_data def __len__(self): return len(self.rgb)
294coder/Efficient-MIF
6,189
task_datasets/WV3.py
import torch import torch.utils.data as data import torchvision.transforms as T import cv2 import numpy as np import h5py from typing import List, Tuple, Optional, Union class Identity: def __call__(self, *args): return args class WV3Datasets(data.Dataset): def __init__( self, file: Union[h5py.File, str, dict], aug_prob=0.0, hp=False, hp_ksize=(5, 5), norm_range=True, full_res=False, ): """ :param d: h5py.File or dict :param aug_prob: augmentation probability :param hp: high pass for ms and pan. x = x - cv2.boxFilter(x) :param hp_ksize: cv2.boxFiler kernel size :param norm_range: normalize data range """ super(WV3Datasets, self).__init__() # FIXME: should pass @path rather than @file which is h5py.File object to avoid can not be pickled error if isinstance(file, (str, h5py.File)): if isinstance(file, str): file = h5py.File(file) print( "warning: when @file is a h5py.File object, it can not be pickled.", "try to set DataLoader number_worker to 0", ) if not full_res: self.gt, self.ms, self.lms, self.pan = self.get_divided(file) print("datasets shape:") print("{:^20}{:^20}{:^20}{:^20}".format("pan", "ms", "lms", "gt")) print( "{:^20}{:^20}{:^20}{:^20}".format( str(tuple(self.pan.shape)), str(tuple(self.ms.shape)), str(tuple(self.lms.shape)), str(tuple(self.gt.shape)), ) ) else: self.ms, self.lms, self.pan = self.get_divided(file, True) print("datasets shape:") print("{:^20}{:^20}{:^20}".format("pan", "ms", "lms")) print( "{:^20}{:^20}{:^20}".format( str(tuple(self.pan.shape)), str(tuple(self.ms.shape)), str(tuple(self.lms.shape)), ) ) self.size = self.ms.shape[0] # highpass filter self.hp = hp self.hp_ksize = hp_ksize if hp and hp_ksize is not None: self.group_high_pass(hp_ksize) # to tensor if norm_range: def norm_func(x): x = x / 2047.0 return x else: def norm_func(x): return x self.pan = norm_func(self.pan) self.ms = norm_func(self.ms) self.lms = norm_func(self.lms) if not full_res: self.gt = norm_func(self.gt) # geometrical transformation self.aug_prob = aug_prob self.geo_trans = ( T.Compose( [T.RandomVerticalFlip(p=aug_prob), T.RandomHorizontalFlip(p=aug_prob)] ) if aug_prob != 0.0 else Identity() ) @staticmethod def get_divided(d, full_resolution=False): if not full_resolution: return ( torch.tensor(d["gt"][:], dtype=torch.float32), torch.tensor(d["ms"][:], dtype=torch.float32), torch.tensor(d["lms"][:], dtype=torch.float32), torch.tensor(d["pan"][:], dtype=torch.float32), ) else: return ( torch.tensor(d["ms"][:], dtype=torch.float32), torch.tensor(d["lms"][:], dtype=torch.float32), torch.tensor(d["pan"][:], dtype=torch.float32), ) @staticmethod def _get_high_pass(data, k_size): for i, img in enumerate(data): hp = cv2.boxFilter(img.transpose(1, 2, 0), -1, k_size) if hp.ndim == 2: hp = hp[..., np.newaxis] data[i] = img - hp.transpose(2, 0, 1) return data def group_high_pass(self, k_size): self.ms = self._get_high_pass(self.ms, k_size) self.pan = self._get_high_pass(self.pan, k_size) def aug_trans(self, *data): data_list = [] seed = torch.random.seed() for d in data: torch.manual_seed(seed) d = self.geo_trans(d) data_list.append(d) return data_list def __getitem__(self, item): if hasattr(self, "gt"): tuple_data = (self.pan[item], self.ms[item], self.lms[item], self.gt[item]) else: tuple_data = (self.pan[item], self.ms[item], self.lms[item]) return self.aug_trans(*tuple_data) if self.aug_prob != 0.0 else tuple_data def __len__(self): return self.size def __repr__(self): return ( f"num: {self.size} \n " f"augmentation: {self.geo_trans} \n" f"get high pass ms and pan: {self.hp} \n " f"filter kernel size: {self.hp_ksize}" ) def make_datasets( path, split_ratio=0.8, hp=True, seed=2022, aug_probs: Tuple = (0.0, 0.0) ): """ if your dataset didn't split before, use this function will split your dataset into two part, which are train and validate datasets. :param device: device :param path: datasets path :param split_ratio: train validate split ratio :param hp: get high pass data, only works for ms and pan data :param seed: split data random state :param aug_probs: augmentation probabilities, type List :return: List[datasets] """ d = h5py.File(path) ds = [ torch.tensor(d["gt"]), torch.tensor(d["ms"]), torch.tensor(d["lms"]), torch.tensor(d["pan"]), ] n = ds[0].shape[0] s = int(n * split_ratio) random_perm = np.arange(n) np.random.seed(seed) np.random.shuffle(random_perm) train_set = {} val_set = {} for i, name in enumerate(["gt", "ms", "lms", "pan"]): ds[i] = ds[i][random_perm] train_set[name] = ds[i][:s] val_set[name] = ds[i][s:] train_ds = WV3Datasets(train_set, hp=hp, aug_prob=aug_probs[0]) val_ds = WV3Datasets(val_set, hp=hp, aug_prob=aug_probs[1]) return train_ds, val_ds
294coder/Efficient-MIF
8,117
utils/optim_utils.py
from typing import Iterable, Optional, Union import weakref import numpy as np import torch import torch.nn as nn import torch.optim as optim from torch.optim.lr_scheduler import ( CosineAnnealingLR, MultiStepLR, CosineAnnealingWarmRestarts, ReduceLROnPlateau, ) import accelerate import torch_ema import deepspeed from deepspeed.runtime.zero import GatheredParameters import sys sys.path.append('./') from utils.misc import is_main_process class IdentityScheduler(torch.optim.lr_scheduler._LRScheduler): # a placeholder for lr_scheduler or weight_decay_scheduler def __init__(self, optim, **kwargs): self.optim = optim self.kwargs = kwargs def step(self, *args, **kwargs): pass def state_dict(self): return self.kwargs def load_state_dict(self, state_dict): pass def cosine_scheduler( base_value, final_value, epochs, niter_per_ep, warmup_epochs=0, start_warmup_value=0 ): """ copy from DINO. manually set learning lr every iteration. note that there is only half epoch of cosine, which means learning rate will not go back to the original. :param base_value: :param final_value: :param epochs: :param niter_per_ep: :param warmup_epochs: :param start_warmup_value: :return: """ warmup_schedule = np.array([]) warmup_iters = warmup_epochs * niter_per_ep if warmup_epochs > 0: warmup_schedule = np.linspace(start_warmup_value, base_value, warmup_iters) iters = np.arange(epochs * niter_per_ep - warmup_iters) schedule = final_value + 0.5 * (base_value - final_value) * ( 1 + np.cos(np.pi * iters / len(iters)) ) schedule = np.concatenate((warmup_schedule, schedule)) assert len(schedule) == epochs * niter_per_ep return schedule class LinearWarmupScheduler: def __init__(self, opt: optim.Optimizer, init_value, warmup_value, warmup_epochs): self.opt = opt self.init_value = init_value self.warmup_value = warmup_value self.warmup_epochs = warmup_epochs self.values = np.linspace(init_value, warmup_value, warmup_epochs) self.now_index = 0 def step(self): self.opt.param_groups[0]["lr"] = self.values[self.now_index] self.now_index += 1 class CosineAnnealingWarmRestartsReduce(CosineAnnealingWarmRestarts): def __init__(self, opt: optim.Optimizer, T_0, T_mult=1, lr_mult=1, eta_min=0, last_epoch=-1): self.opt = opt self.lr_mult = lr_mult super().__init__(opt, T_0, T_mult, eta_min, last_epoch) def step(self, epoch=None): super().step(epoch) if self.T_cur == self.T_i-1 and self.last_epoch != 0: # reduce the base lr for i in range(len(self.base_lrs)): self.base_lrs[i] *= self.lr_mult self.base_lrs[i] = max(self.base_lrs[i], self.eta_min) def get_precision(mixed_precision): if mixed_precision == 'fp32' or mixed_precision == 'no': return torch.float32 elif mixed_precision == 'fp16': return torch.float16 elif mixed_precision == 'bf16': return torch.bfloat16 else: raise ValueError(f"Invalid mixed precision value: {mixed_precision}") def get_scheduler(optim, **kwargs): """ get lr_scheduler or weight_decay_scheduler Args: optim: optimizer **kwargs: a dict containing type of scheduler and its arguments Returns: a scheduler """ name = kwargs["name"] kwargs.pop("name") if name == "cos_anneal": return CosineAnnealingLR(optim, **kwargs) elif name == "cos_anneal_restart": return CosineAnnealingWarmRestarts(optim, **kwargs) elif name == "cos_anneal_restart_reduce": return CosineAnnealingWarmRestartsReduce(optim, **kwargs) elif name == "multi_step": return MultiStepLR(optim, **kwargs) elif name == "plateau": return ReduceLROnPlateau(optim, **kwargs) elif name == "identity": return IdentityScheduler(optim, **kwargs) else: raise NotImplementedError def get_optimizer(model: torch.nn.Module, params: "Iterable | dict", **kwargs): name = kwargs["name"] kwargs.pop("name") if name == "sgd": return optim.SGD(params, **kwargs) elif name == "adam": return optim.Adam(params, **kwargs) elif name == "adamw": return optim.AdamW(params, **kwargs) elif name == 'lion': from lion_pytorch import Lion return Lion(params, betas=(0.95, 0.98), use_triton=True, **kwargs) elif name == 'fusedadam': return deepspeed.ops.adam.FusedAdam(params, **kwargs) elif name == 'schedulefree-adam': import schedulefree return schedulefree.AdamWScheduleFree(params, **kwargs) elif name == 'adam-mini': return Adam_mini(model, **kwargs) else: raise NotImplementedError(f'optimizer {name} not implemented') def get_ema_model(parameters: Iterable[torch.nn.Parameter], accelerator: accelerate.Accelerator=None, **ema_kwargs): if accelerator is not None: if accelerator.state.deepspeed_plugin is not None: ema_model = DeepspeedEMA(parameters, **ema_kwargs) return ema_model ema_model = torch_ema.ema.ExponentialMovingAverage(model, **ema_kwargs) return ema_model if __name__ == "__main__": import matplotlib.pyplot as plt import torch.optim as optim import torch.nn as nn # init_lr = 1e-3 # final_lr = 1e-6 # epochs = 500 # # nither_per_ep = int(np.ceil(3000 // 16)) # len(datasets) / batch_size # # warm_epochs = 80 # # start_warmup_value = init_lr # # cos_sche = cosine_scheduler( # # init_lr, final_lr, epochs, nither_per_ep, warm_epochs, start_warmup_value # # ) # # plt.plot(list(map(lambda x: x / nither_per_ep, range(len(cos_sche)))), cos_sche) # # plt.show() # # torch cosine annealing lr scheduler # net = nn.Sequential(nn.Linear(8, 64)) # optimizer = optim.AdamW(net.parameters(), lr=init_lr) # # cos_sche2 = CosineAnnealingLR(optimizer, epochs - warm_epochs, final_lr) # cos_anneal_reduce_sche = CosineAnnealingWarmRestartsReduce(optimizer, 50, 2, 0.5, 1e-6, last_epoch=-1) # lr = [] # for i in range(200, 500): # l = optimizer.param_groups[0]["lr"] # lr.append(l) # # if i > warm_epochs: # # cos_sche2.step() # cos_anneal_reduce_sche.step(i) # plt.plot(range(200, 500), lr) # # plt.show() # plt.savefig('cos_anneal_reduce.png') import accelerate from torch.utils.data import DataLoader, Dataset, TensorDataset accelerator = accelerate.Accelerator() model = torch.nn.Linear(3, 16) opt = optim.Adam(model.parameters(), lr=1e-3) dl = DataLoader(TensorDataset(*([torch.randn(16, 3)]*2)), batch_size=1) model, dl, opt = accelerator.prepare(model, dl, opt) ema_model = get_ema_model(list(model.parameters()), accelerator=accelerator, decay=0.9) accelerator.wait_for_everyone() if is_main_process(): print('main process: set weight to zero weight \n') model.weight.data.zero_() model.bias.data.zero_() print(model.weight) else: import time time.sleep(2) # the other process print('other process: set weight to non-zero weight\n') print(ema_model.shadow_params) accelerator.wait_for_everyone() print('-----------------------------'*2, '\n') ema_model.update() # proc 0 is zeros but proc 1 is not, # so the ema_model should not be all zeros if is_main_process(): print('main process') print(ema_model.shadow_params) print(model.weight) print('-----------------------------'*2, '\n') ema_model.restore(model.parameters()) print(model.weight) print('-----------------------------'*2)
2977094657/DidaAPI
4,920
frontend/docs/.vitepress/theme/style/vars.css
/** * Customize default theme styling by overriding CSS variables: * https://github.com/vuejs/vitepress/blob/main/src/client/theme-default/styles/vars.css */ /** * Colors * * Each colors have exact same color scale system with 3 levels of solid * colors with different brightness, and 1 soft color. * * - `XXX-1`: The most solid color used mainly for colored text. It must * satisfy the contrast ratio against when used on top of `XXX-soft`. * * - `XXX-2`: The color used mainly for hover state of the button. * * - `XXX-3`: The color for solid background, such as bg color of the button. * It must satisfy the contrast ratio with pure white (#ffffff) text on * top of it. * * - `XXX-soft`: The color used for subtle background such as custom container * or badges. It must satisfy the contrast ratio when putting `XXX-1` colors * on top of it. * * The soft color must be semi transparent alpha channel. This is crucial * because it allows adding multiple "soft" colors on top of each other * to create a accent, such as when having inline code block inside * custom containers. * * - `default`: The color used purely for subtle indication without any * special meanings attched to it such as bg color for menu hover state. * * - `brand`: Used for primary brand colors, such as link text, button with * brand theme, etc. * * - `tip`: Used to indicate useful information. The default theme uses the * brand color for this by default. * * - `warning`: Used to indicate warning to the users. Used in custom * container, badges, etc. * * - `danger`: Used to indicate dangerous message to the users. Used in custom * container, badges, etc. * -------------------------------------------------------------------------- */ :root { --vp-c-default-1: var(--vp-c-gray-1); --vp-c-default-2: var(--vp-c-gray-2); --vp-c-default-3: var(--vp-c-gray-3); --vp-c-default-soft: var(--vp-c-gray-soft); /* 使用滴答清单官方色彩 */ --vp-c-brand-1: #4772FA; --vp-c-brand-2: #5a7efb; --vp-c-brand-3: #6d8afc; --vp-c-brand-soft: rgba(71, 114, 250, 0.14); --vp-c-tip-1: var(--vp-c-brand-1); --vp-c-tip-2: var(--vp-c-brand-2); --vp-c-tip-3: var(--vp-c-brand-3); --vp-c-tip-soft: var(--vp-c-brand-soft); /* 使用滴答清单的橙色作为警告色 */ --vp-c-warning-1: #FFB000; --vp-c-warning-2: #ffba1a; --vp-c-warning-3: #ffc433; --vp-c-warning-soft: rgba(255, 176, 0, 0.14); --vp-c-danger-1: var(--vp-c-red-1); --vp-c-danger-2: var(--vp-c-red-2); --vp-c-danger-3: var(--vp-c-red-3); --vp-c-danger-soft: var(--vp-c-red-soft); } /** * Component: Button * -------------------------------------------------------------------------- */ :root { --vp-button-brand-border: transparent; --vp-button-brand-text: var(--vp-c-white); --vp-button-brand-bg: var(--vp-c-brand-3); --vp-button-brand-hover-border: transparent; --vp-button-brand-hover-text: var(--vp-c-white); --vp-button-brand-hover-bg: var(--vp-c-brand-2); --vp-button-brand-active-border: transparent; --vp-button-brand-active-text: var(--vp-c-white); --vp-button-brand-active-bg: var(--vp-c-brand-1); } /** * Component: Home * -------------------------------------------------------------------------- */ :root { --vp-home-hero-name-color: transparent; --vp-home-hero-name-background: -webkit-linear-gradient( 120deg, #4772FA 30%, #FFB000 ); --vp-home-hero-image-background-image: linear-gradient( -45deg, #4772FA 50%, #FFB000 50% ); --vp-home-hero-image-filter: blur(44px); } @media (min-width: 640px) { :root { --vp-home-hero-image-filter: blur(56px); } } @media (min-width: 960px) { :root { --vp-home-hero-image-filter: blur(68px); } } /** * Component: Custom Block * -------------------------------------------------------------------------- */ :root { --vp-custom-block-tip-border: transparent; --vp-custom-block-tip-text: var(--vp-c-text-1); --vp-custom-block-tip-bg: var(--vp-c-brand-soft); --vp-custom-block-tip-code-bg: var(--vp-c-brand-soft); } /** * Component: Algolia * -------------------------------------------------------------------------- */ .DocSearch { --docsearch-primary-color: var(--vp-c-brand-1) !important; } /** * Component: Layout adjustments * -------------------------------------------------------------------------- */ /* 增加首页顶部距离 */ .VPHome { padding-top: 40px !important; } /* 调整Hero区域间距 */ .VPHero .container { padding-top: 30px !important; padding-bottom: 40px !important; } /* 调整Features区域间距 */ .VPFeatures { padding-top: 40px !important; padding-bottom: 80px !important; } /* 响应式调整 */ @media (max-width: 768px) { .VPHome { padding-top: 40px !important; } .VPHero .container { padding-top: 40px !important; padding-bottom: 40px !important; } .VPFeatures { padding-top: 20px !important; padding-bottom: 40px !important; } }
2977094657/DidaAPI
2,468
frontend/docs/api/habits/week-current-statistics.md
# 获取本周习惯打卡统计 ## 接口信息 - **接口路径**: `GET /habits/statistics/week/current` - **接口描述**: 获取本周的习惯打卡统计信息 - **请求方式**: GET - **认证要求**: 需要先完成微信登录获取认证会话 ## 请求参数 无需参数 ## 请求示例 ```bash curl -X GET "http://localhost:8000/habits/statistics/week/current" ``` ## 响应格式 ### 成功响应 ```json { "weekStart": "2023-12-04", "weekEnd": "2023-12-10", "totalHabits": 5, "completedDays": 4, "completionRate": 80.0, "dailyStats": [ { "date": "2023-12-04", "dayOfWeek": "Monday", "completedHabits": 4, "totalHabits": 5, "completionRate": 80.0 }, { "date": "2023-12-05", "dayOfWeek": "Tuesday", "completedHabits": 5, "totalHabits": 5, "completionRate": 100.0 } ], "habitDetails": [ { "habitId": "habit123", "habitName": "早起", "targetDays": 7, "completedDays": 5, "completionRate": 71.4, "streak": 3, "dailyStatus": [ { "date": "2023-12-04", "completed": true }, { "date": "2023-12-05", "completed": true } ] } ] } ``` ### 错误响应 #### 未认证 ```json { "error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录" } ``` #### 服务错误 ```json { "error": "service_error", "message": "获取本周习惯打卡统计失败,请稍后重试" } ``` ## 响应字段说明 | 字段名 | 类型 | 描述 | |--------|------|------| | weekStart | string | 本周开始日期 | | weekEnd | string | 本周结束日期 | | totalHabits | integer | 总习惯数量 | | completedDays | integer | 完成打卡的天数 | | completionRate | float | 整体完成率(百分比) | | dailyStats | array | 每日统计数据 | | dailyStats[].date | string | 日期 | | dailyStats[].dayOfWeek | string | 星期几 | | dailyStats[].completedHabits | integer | 当日完成的习惯数 | | dailyStats[].totalHabits | integer | 当日总习惯数 | | dailyStats[].completionRate | float | 当日完成率 | | habitDetails | array | 习惯详细信息 | | habitDetails[].habitId | string | 习惯ID | | habitDetails[].habitName | string | 习惯名称 | | habitDetails[].targetDays | integer | 目标天数 | | habitDetails[].completedDays | integer | 已完成天数 | | habitDetails[].completionRate | float | 习惯完成率 | | habitDetails[].streak | integer | 连续打卡天数 | | habitDetails[].dailyStatus | array | 每日打卡状态 | ## 使用说明 1. **认证要求**: 需要先调用微信登录接口获取认证会话 2. **统计周期**: 统计当前自然周(周一到周日)的数据 3. **实时更新**: 数据实时更新,反映最新的打卡状态 4. **多维度统计**: 提供整体、每日、每个习惯的多维度统计 ## 相关接口 - [获取所有习惯](../habits.md) - [导出习惯数据](./export-habits.md) - [微信登录流程](../auth/wechat-login-flow.md) ## 注意事项 - 需要先完成微信登录获取认证会话 - 统计数据基于当前自然周 - 完成率保留一位小数 - 连续打卡天数会在中断后重新计算
281677160/openwrt-package
72,103
luci-app-passwall/root/usr/share/passwall/iptables.sh
#!/bin/sh DIR="$(cd "$(dirname "$0")" && pwd)" MY_PATH=$DIR/iptables.sh IPSET_LOCAL="passwall_local" IPSET_LAN="passwall_lan" IPSET_VPS="passwall_vps" IPSET_SHUNT="passwall_shunt" IPSET_GFW="passwall_gfw" IPSET_CHN="passwall_chn" IPSET_BLACK="passwall_black" IPSET_WHITE="passwall_white" IPSET_BLOCK="passwall_block" IPSET_LOCAL6="passwall_local6" IPSET_LAN6="passwall_lan6" IPSET_VPS6="passwall_vps6" IPSET_SHUNT6="passwall_shunt6" IPSET_GFW6="passwall_gfw6" IPSET_CHN6="passwall_chn6" IPSET_BLACK6="passwall_black6" IPSET_WHITE6="passwall_white6" IPSET_BLOCK6="passwall_block6" FORCE_INDEX=2 USE_SHUNT_TCP=0 USE_SHUNT_UDP=0 . /lib/functions/network.sh ipt=$(command -v iptables-legacy || command -v iptables) ip6t=$(command -v ip6tables-legacy || command -v ip6tables) ipt_n="$ipt -t nat -w" ipt_m="$ipt -t mangle -w" ip6t_n="$ip6t -t nat -w" ip6t_m="$ip6t -t mangle -w" [ -z "$ip6t" -o -z "$(lsmod | grep 'ip6table_nat')" ] && ip6t_n="eval #$ip6t_n" [ -z "$ip6t" -o -z "$(lsmod | grep 'ip6table_mangle')" ] && ip6t_m="eval #$ip6t_m" FWI=$(uci -q get firewall.passwall.path 2>/dev/null) FAKE_IP="198.18.0.0/15" FAKE_IP_6="fc00::/18" factor() { if [ -z "$1" ] || [ -z "$2" ]; then echo "" elif [ "$1" == "1:65535" ]; then echo "" else echo "$2 $1" fi } dst() { echo "-m set $2 --match-set $1 dst" } comment() { local name=$(echo $1 | sed 's/ /_/g') echo "-m comment --comment '$name'" } #解决端口超过15个ipt无效,支持单端口、端口范围 add_port_rules() { local ipt_cmd="$1" local port_list="$2" local target="$3" echo "$port_list" | grep -vq '[0-9]' && return port_list=$(echo "$port_list" | tr -d ' ' | sed 's/-/:/g' | tr ',' '\n' | awk '!a[$0]++' | grep -v '^$') [ -z "$port_list" ] && return if echo "$port_list" | grep -q '^1:65535$'; then eval "$ipt_cmd $target" return fi local multiport_ports="" local range_ports="" local count=0 local port for port in $port_list; do if echo "$port" | grep -q ':'; then range_ports="$range_ports $port" else multiport_ports="$multiport_ports,$port" count=$((count + 1)) if [ "$count" -eq 15 ]; then eval "$ipt_cmd -m multiport --dport ${multiport_ports#,} $target" multiport_ports="" count=0 fi fi done if [ -n "$multiport_ports" ]; then eval "$ipt_cmd -m multiport --dport ${multiport_ports#,} $target" fi for port in $range_ports; do eval "$ipt_cmd --dport $port $target" done } destroy_ipset() { for i in "$@"; do ipset -q -F $i ipset -q -X $i done } insert_rule_before() { [ $# -ge 3 ] || { return 1 } local ipt_tmp="${1}"; shift local chain="${1}"; shift local keyword="${1}"; shift local rule="${1}"; shift local default_index="${1}"; shift default_index=${default_index:-0} local _index=$($ipt_tmp -n -L $chain --line-numbers 2>/dev/null | grep "$keyword" | head -n 1 | awk '{print $1}') if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then $ipt_tmp -A $chain $rule else if [ -z "${_index}" ]; then _index=${default_index} fi $ipt_tmp -I $chain $_index $rule fi } insert_rule_after() { [ $# -ge 3 ] || { return 1 } local ipt_tmp="${1}"; shift local chain="${1}"; shift local keyword="${1}"; shift local rule="${1}"; shift local default_index="${1}"; shift default_index=${default_index:-0} local _index=$($ipt_tmp -n -L $chain --line-numbers 2>/dev/null | grep "$keyword" | awk 'END {print}' | awk '{print $1}') if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then $ipt_tmp -A $chain $rule else if [ -n "${_index}" ]; then _index=$((_index + 1)) else _index=${default_index} fi $ipt_tmp -I $chain $_index $rule fi } RULE_LAST_INDEX() { [ $# -ge 3 ] || { echolog "索引列举方式不正确(iptables),终止执行!" return 1 } local ipt_tmp="${1}"; shift local chain="${1}"; shift local list="${1}"; shift local default="${1:-0}"; shift local _index=$($ipt_tmp -n -L $chain --line-numbers 2>/dev/null | grep "$list" | head -n 1 | awk '{print $1}') echo "${_index:-${default}}" } REDIRECT() { local s="-j REDIRECT" if [ -n "$1" ]; then s="$s --to-ports $1" if [ "$2" = "MARK" ]; then s="-j MARK --set-mark $1" elif [ "$2" = "TPROXY" ]; then local mark="-m mark --mark 1" s="${mark} -j TPROXY --tproxy-mark 1/1 --on-port $1" fi fi echo "$s" } get_jump_ipt() { case "$1" in direct) local mark="-m mark ! --mark 1" s="${mark} -j RETURN" echo $s ;; proxy) if [ -n "$2" ] && [ -n "$(echo $2 | grep "^-")" ]; then echo "$2" else echo "$(REDIRECT $2 $3)" fi ;; esac } gen_lanlist() { cat $RULES_PATH/lanlist_ipv4 | tr -s '\n' | grep -v "^#" } gen_lanlist_6() { cat $RULES_PATH/lanlist_ipv6 | tr -s '\n' | grep -v "^#" } get_wan_ip() { local NET_IF local NET_ADDR network_flush_cache network_find_wan NET_IF network_get_ipaddr NET_ADDR "${NET_IF}" echo $NET_ADDR } get_wan6_ip() { local NET_IF local NET_ADDR network_flush_cache network_find_wan6 NET_IF network_get_ipaddr6 NET_ADDR "${NET_IF}" echo $NET_ADDR } load_acl() { ([ "$ENABLED_ACLS" == 1 ] || ([ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ])) && echolog " - 访问控制:" [ "$ENABLED_ACLS" == 1 ] && { acl_app for sid in $(ls -F ${TMP_ACL_PATH} | grep '/$' | awk -F '/' '{print $1}' | grep -v 'default'); do eval $(uci -q show "${CONFIG}.${sid}" | cut -d'.' -sf 3-) tcp_no_redir_ports=${tcp_no_redir_ports:-default} udp_no_redir_ports=${udp_no_redir_ports:-default} use_global_config=${use_global_config:-0} tcp_proxy_drop_ports=${tcp_proxy_drop_ports:-default} udp_proxy_drop_ports=${udp_proxy_drop_ports:-default} tcp_redir_ports=${tcp_redir_ports:-default} udp_redir_ports=${udp_redir_ports:-default} use_direct_list=${use_direct_list:-1} use_proxy_list=${use_proxy_list:-1} use_block_list=${use_block_list:-1} use_gfw_list=${use_gfw_list:-1} chn_list=${chn_list:-direct} tcp_proxy_mode=${tcp_proxy_mode:-proxy} udp_proxy_mode=${udp_proxy_mode:-proxy} [ "$tcp_no_redir_ports" = "default" ] && tcp_no_redir_ports=$TCP_NO_REDIR_PORTS [ "$udp_no_redir_ports" = "default" ] && udp_no_redir_ports=$UDP_NO_REDIR_PORTS [ "$tcp_proxy_drop_ports" = "default" ] && tcp_proxy_drop_ports=$TCP_PROXY_DROP_PORTS [ "$udp_proxy_drop_ports" = "default" ] && udp_proxy_drop_ports=$UDP_PROXY_DROP_PORTS [ "$tcp_redir_ports" = "default" ] && tcp_redir_ports=$TCP_REDIR_PORTS [ "$udp_redir_ports" = "default" ] && udp_redir_ports=$UDP_REDIR_PORTS [ -n "$(get_cache_var "ACL_${sid}_tcp_node")" ] && tcp_node=$(get_cache_var "ACL_${sid}_tcp_node") [ -n "$(get_cache_var "ACL_${sid}_tcp_redir_port")" ] && tcp_port=$(get_cache_var "ACL_${sid}_tcp_redir_port") [ -n "$(get_cache_var "ACL_${sid}_udp_node")" ] && udp_node=$(get_cache_var "ACL_${sid}_udp_node") [ -n "$(get_cache_var "ACL_${sid}_udp_redir_port")" ] && udp_port=$(get_cache_var "ACL_${sid}_udp_redir_port") [ -n "$(get_cache_var "ACL_${sid}_dns_port")" ] && dns_redirect_port=$(get_cache_var "ACL_${sid}_dns_port") [ -n "$tcp_node" ] && tcp_node_remark=$(config_n_get $tcp_node remarks) [ -n "$udp_node" ] && udp_node_remark=$(config_n_get $udp_node remarks) use_shunt_tcp=0 use_shunt_udp=0 [ -n "$tcp_node" ] && [ "$(config_n_get $tcp_node protocol)" = "_shunt" ] && use_shunt_tcp=1 [ -n "$udp_node" ] && [ "$(config_n_get $udp_node protocol)" = "_shunt" ] && use_shunt_udp=1 [ "${use_global_config}" = "1" ] && { tcp_node_remark=$(config_n_get $TCP_NODE remarks) udp_node_remark=$(config_n_get $UDP_NODE remarks) use_direct_list=${USE_DIRECT_LIST} use_proxy_list=${USE_PROXY_LIST} use_block_list=${USE_BLOCK_LIST} use_gfw_list=${USE_GFW_LIST} chn_list=${CHN_LIST} tcp_proxy_mode=${TCP_PROXY_MODE} udp_proxy_mode=${UDP_PROXY_MODE} use_shunt_tcp=${USE_SHUNT_TCP} use_shunt_udp=${USE_SHUNT_UDP} dns_redirect_port=${DNS_REDIRECT_PORT} black_set_name=${IPSET_BLACK} black6_set_name=${IPSET_BLACK6} gfw_set_name=${IPSET_GFW} gfw6_set_name=${IPSET_GFW6} shunt_set_name=${IPSET_SHUNT} shunt6_set_name=${IPSET_SHUNT6} } _acl_list=${TMP_ACL_PATH}/${sid}/source_list for i in $(cat $_acl_list); do local _ipt_source _ipv4 local msg if [ -n "${interface}" ]; then . /lib/functions/network.sh local gateway device network_get_gateway gateway "${interface}" network_get_device device "${interface}" [ -z "${device}" ] && device="${interface}" _ipt_source="-i ${device} " msg="源接口【${device}】," else msg="源接口【所有】," fi if [ -n "$(echo ${i} | grep '^iprange:')" ]; then _iprange=$(echo ${i} | sed 's#iprange:##g') _ipt_source=$(factor ${_iprange} "${_ipt_source}-m iprange --src-range") msg="${msg}IP range【${_iprange}】," _ipv4="1" unset _iprange elif [ -n "$(echo ${i} | grep '^ipset:')" ]; then _ipset=$(echo ${i} | sed 's#ipset:##g') msg="${msg}IPset【${_ipset}】," ipset -q list ${_ipset} >/dev/null if [ $? -eq 0 ]; then _ipt_source="${_ipt_source}-m set --match-set ${_ipset} src" unset _ipset else echolog " - 【$remarks】,${msg}不存在,忽略。" unset _ipset continue fi elif [ -n "$(echo ${i} | grep '^ip:')" ]; then _ip=$(echo ${i} | sed 's#ip:##g') _ipt_source=$(factor ${_ip} "${_ipt_source}-s") msg="${msg}IP【${_ip}】," _ipv4="1" unset _ip elif [ -n "$(echo ${i} | grep '^mac:')" ]; then _mac=$(echo ${i} | sed 's#mac:##g') _ipt_source=$(factor ${_mac} "${_ipt_source}-m mac --mac-source") msg="${msg}MAC【${_mac}】," unset _mac elif [ -n "$(echo ${i} | grep '^any')" ]; then msg="${msg}所有设备," else continue fi msg="【$remarks】,${msg}" ipt_tmp=$ipt_n [ -n "${is_tproxy}" ] && ipt_tmp=$ipt_m [ "$tcp_no_redir_ports" != "disable" ] && { if ! has_1_65535 "$tcp_no_redir_ports"; then [ "$_ipv4" != "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p tcp" $tcp_no_redir_ports "-j RETURN" 2>/dev/null add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") ${_ipt_source} -p tcp" $tcp_no_redir_ports "-j RETURN" echolog " - ${msg}不代理 TCP 端口[${tcp_no_redir_ports}]" else #结束时会return,无需加多余的规则。 unset tcp_port echolog " - ${msg}不代理所有 TCP 端口" fi } [ "$udp_no_redir_ports" != "disable" ] && { if ! has_1_65535 "$udp_no_redir_ports"; then [ "$_ipv4" != "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp" $udp_no_redir_ports "-j RETURN" 2>/dev/null add_port_rules "$ipt_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp" $udp_no_redir_ports "-j RETURN" echolog " - ${msg}不代理 UDP 端口[${udp_no_redir_ports}]" else #结束时会return,无需加多余的规则。 unset udp_port echolog " - ${msg}不代理所有 UDP 端口" fi } local dns_redirect [ $(config_t_get global dns_redirect "1") = "1" ] && dns_redirect=53 if ([ -n "$tcp_port" ] && [ -n "${tcp_proxy_mode}" ]) || ([ -n "$udp_port" ] && [ -n "${udp_proxy_mode}" ]); then [ "${use_proxy_list}" = "1" ] && { [ "${use_global_config}" = "0" ] && { black_set_name="passwall_${sid}_black" black6_set_name="passwall_${sid}_black6" ipset -! create $black_set_name nethash maxelem 1048576 timeout 172800 ipset -! create $black6_set_name nethash family inet6 maxelem 1048576 timeout 172800 } } [ "${use_gfw_list}" = "1" ] && { [ "${use_global_config}" = "0" ] && { gfw_set_name="passwall_${sid}_gfw" gfw6_set_name="passwall_${sid}_gfw6" ipset -! create $gfw_set_name nethash maxelem 1048576 timeout 172800 ipset -! create $gfw6_set_name nethash family inet6 maxelem 1048576 timeout 172800 } } [ "${use_shunt_tcp}" = "1" -o "${use_shunt_udp}" = "1" ] && { [ "${use_global_config}" = "0" ] && { shunt_set_name="passwall_${sid}_shunt" shunt6_set_name="passwall_${sid}_shunt6" ipset -! create $shunt_set_name nethash maxelem 1048576 timeout 172800 ipset -! create $shunt6_set_name nethash family inet6 maxelem 1048576 timeout 172800 } } [ -n "${dns_redirect_port}" ] && dns_redirect=${dns_redirect_port} else [ -n "${DIRECT_DNSMASQ_PORT}" ] && dns_redirect=${DIRECT_DNSMASQ_PORT} fi if [ -n "${dns_redirect}" ]; then $ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j RETURN [ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j RETURN 2>/dev/null $ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j RETURN [ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j RETURN 2>/dev/null $ipt_n -A PSW_DNS $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect} [ "$_ipv4" != "1" ] && $ip6t_n -A PSW_DNS $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect} 2>/dev/null $ipt_n -A PSW_DNS $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect} [ "$_ipv4" != "1" ] && $ip6t_n -A PSW_DNS $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect} 2>/dev/null [ -z "$(get_cache_var "ACL_${sid}_tcp_default")" ] && echolog " - ${msg}使用与全局配置不相同节点,已将DNS强制重定向到专用 DNS 服务器。" fi [ -n "$tcp_port" -o -n "$udp_port" ] && { [ "${use_block_list}" = "1" ] && $ipt_m -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_BLOCK) -j DROP [ "${use_direct_list}" = "1" ] && $ipt_tmp -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_WHITE) -j RETURN [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { [ "${use_block_list}" = "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_BLOCK6) -j DROP 2>/dev/null [ "${use_direct_list}" = "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_WHITE6) -j RETURN 2>/dev/null } [ "$tcp_proxy_drop_ports" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-d $FAKE_IP_6 -j DROP" 2>/dev/null [ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $black6_set_name) -j DROP" 2>/dev/null [ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $gfw6_set_name) -j DROP" 2>/dev/null [ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j DROP")" 2>/dev/null [ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $shunt6_set_name) -j DROP" 2>/dev/null [ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-j DROP" 2>/dev/null } add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-d $FAKE_IP -j DROP" [ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $black_set_name) -j DROP" [ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $gfw_set_name) -j DROP" [ "${chn_list}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "-j DROP")" [ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $shunt_set_name) -j DROP" [ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-j DROP" echolog " - ${msg}屏蔽代理 TCP 端口[${tcp_proxy_drop_ports}]" } [ "$udp_proxy_drop_ports" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-d $FAKE_IP_6 -j DROP" 2>/dev/null [ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $black6_set_name) -j DROP" 2>/dev/null [ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $gfw6_set_name) -j DROP" 2>/dev/null [ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j DROP")" 2>/dev/null [ "${use_shunt_udp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $shunt6_set_name) -j DROP" 2>/dev/null [ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-j DROP" 2>/dev/null } add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-d $FAKE_IP -j DROP" [ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $black_set_name) -j DROP" [ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $gfw_set_name) -j DROP" [ "${chn_list}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "-j DROP")" [ "${use_shunt_udp}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $shunt_set_name) -j DROP" [ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-j DROP" echolog " - ${msg}屏蔽代理 UDP 端口[${udp_proxy_drop_ports}]" } } [ -n "$tcp_port" ] && { if [ -n "${tcp_proxy_mode}" ]; then msg2="${msg}使用 TCP 节点[$tcp_node_remark]" if [ -n "${is_tproxy}" ]; then msg2="${msg2}(TPROXY:${tcp_port})" ipt_tmp=$ipt_m ipt_j="-j PSW_RULE" else msg2="${msg2}(REDIRECT:${tcp_port})" ipt_j="$(REDIRECT $tcp_port)" fi [ "$accept_icmp" = "1" ] && { $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} -d $FAKE_IP $(REDIRECT) [ "${use_proxy_list}" = "1" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $black_set_name) $(REDIRECT) [ "${use_gfw_list}" = "1" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $gfw_set_name) $(REDIRECT) [ "${chn_list}" != "0" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $IPSET_CHN) $(get_jump_ipt ${chn_list}) [ "${use_shunt_tcp}" = "1" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $shunt_set_name) $(REDIRECT) [ "${tcp_proxy_mode}" != "disable" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(REDIRECT) } [ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} -d $FAKE_IP_6 $(REDIRECT) 2>/dev/null [ "${use_proxy_list}" = "1" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $black6_set_name) $(REDIRECT) 2>/dev/null [ "${use_gfw_list}" = "1" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $gfw6_set_name) $(REDIRECT) 2>/dev/null [ "${chn_list}" != "0" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list}) 2>/dev/null [ "${use_shunt_tcp}" = "1" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $shunt6_set_name) $(REDIRECT) 2>/dev/null [ "${tcp_proxy_mode}" != "disable" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(REDIRECT) 2>/dev/null } $ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source} -d $FAKE_IP ${ipt_j} [ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $black_set_name) ${ipt_j}" [ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $gfw_set_name) ${ipt_j}" [ "${chn_list}" != "0" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "${ipt_j}")" [ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $shunt_set_name) ${ipt_j}" [ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "${ipt_j}" [ -n "${is_tproxy}" ] && $ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY) [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { $ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} -d $FAKE_IP_6 -j PSW_RULE 2>/dev/null [ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $black6_set_name) -j PSW_RULE" 2>/dev/null [ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $gfw6_set_name) -j PSW_RULE" 2>/dev/null [ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j PSW_RULE")" 2>/dev/null [ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $shunt6_set_name) -j PSW_RULE" 2>/dev/null [ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "-j PSW_RULE" 2>/dev/null $ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY) 2>/dev/null } else msg2="${msg}不代理 TCP" fi echolog " - ${msg2}" } [ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p tcp -j RETURN 2>/dev/null $ipt_tmp -A PSW $(comment "$remarks") ${_ipt_source} -p tcp -j RETURN [ -n "$udp_port" ] && { if [ -n "${udp_proxy_mode}" ]; then msg2="${msg}使用 UDP 节点[$udp_node_remark]" msg2="${msg2}(TPROXY:${udp_port})" $ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} -d $FAKE_IP -j PSW_RULE [ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $black_set_name) -j PSW_RULE" [ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $gfw_set_name) -j PSW_RULE" [ "${chn_list}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "-j PSW_RULE")" [ "${use_shunt_udp}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $shunt_set_name) -j PSW_RULE" [ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "-j PSW_RULE" $ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} $(REDIRECT $udp_port TPROXY) [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && { $ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} -d $FAKE_IP_6 -j PSW_RULE 2>/dev/null [ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $black6_set_name) -j PSW_RULE" 2>/dev/null [ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $gfw6_set_name) -j PSW_RULE" 2>/dev/null [ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j PSW_RULE")" 2>/dev/null [ "${use_shunt_udp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $shunt6_set_name) -j PSW_RULE" 2>/dev/null [ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "-j PSW_RULE" 2>/dev/null $ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} $(REDIRECT $udp_port TPROXY) 2>/dev/null } else msg2="${msg}不代理 UDP" fi echolog " - ${msg2}" } [ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp -j RETURN 2>/dev/null $ipt_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp -j RETURN unset ipt_tmp ipt_j _ipt_source msg msg2 _ipv4 done unset enabled sid remarks sources use_global_config use_direct_list use_proxy_list use_block_list use_gfw_list chn_list tcp_proxy_mode udp_proxy_mode dns_redirect_port tcp_no_redir_ports udp_no_redir_ports tcp_proxy_drop_ports udp_proxy_drop_ports tcp_redir_ports udp_redir_ports tcp_node udp_node interface unset tcp_port udp_port tcp_node_remark udp_node_remark _acl_list use_shunt_tcp use_shunt_udp dns_redirect done } [ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ] && { msg="【默认】," local ipt_tmp=$ipt_n [ -n "${is_tproxy}" ] && ipt_tmp=$ipt_m [ "$TCP_NO_REDIR_PORTS" != "disable" ] && { add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN" add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN" if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]" else unset TCP_PROXY_MODE echolog " - ${msg}不代理所有 TCP 端口" fi } [ "$UDP_NO_REDIR_PORTS" != "disable" ] && { add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_NO_REDIR_PORTS "-j RETURN" add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_NO_REDIR_PORTS "-j RETURN" if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]" else unset UDP_PROXY_MODE echolog " - ${msg}不代理所有 UDP 端口" fi } local DNS_REDIRECT [ $(config_t_get global dns_redirect "1") = "1" ] && DNS_REDIRECT=53 if ([ -n "$TCP_NODE" ] && [ -n "${TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${UDP_PROXY_MODE}" ]); then [ -n "${DNS_REDIRECT_PORT}" ] && DNS_REDIRECT=${DNS_REDIRECT_PORT} else [ -n "${DIRECT_DNSMASQ_PORT}" ] && DNS_REDIRECT=${DIRECT_DNSMASQ_PORT} fi if [ -n "${DNS_REDIRECT}" ]; then $ipt_m -A PSW $(comment "默认") -p udp --dport 53 -j RETURN $ip6t_m -A PSW $(comment "默认") -p udp --dport 53 -j RETURN 2>/dev/null $ipt_m -A PSW $(comment "默认") -p tcp --dport 53 -j RETURN $ip6t_m -A PSW $(comment "默认") -p tcp --dport 53 -j RETURN 2>/dev/null $ipt_n -A PSW_DNS $(comment "默认") -p udp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT} $ip6t_n -A PSW_DNS $(comment "默认") -p udp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT} 2>/dev/null $ipt_n -A PSW_DNS $(comment "默认") -p tcp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT} $ip6t_n -A PSW_DNS $(comment "默认") -p tcp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT} 2>/dev/null fi [ -n "${TCP_PROXY_MODE}" -o -n "${UDP_PROXY_MODE}" ] && { [ "${USE_BLOCK_LIST}" = "1" ] && $ipt_m -A PSW $(comment "默认") $(dst $IPSET_BLOCK) -j DROP [ "${USE_DIRECT_LIST}" = "1" ] && $ipt_tmp -A PSW $(comment "默认") $(dst $IPSET_WHITE) -j RETURN [ "$PROXY_IPV6" == "1" ] && { [ "${USE_BLOCK_LIST}" = "1" ] && $ip6t_m -A PSW $(comment "默认") $(dst $IPSET_BLOCK6) -j DROP 2>/dev/null [ "${USE_DIRECT_LIST}" = "1" ] && $ip6t_m -A PSW $(comment "默认") $(dst $IPSET_WHITE6) -j RETURN 2>/dev/null } [ "$TCP_PROXY_DROP_PORTS" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && { add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "-d $FAKE_IP_6 -j DROP" [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK6) -j DROP" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_GFW6) -j DROP" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j DROP")" [ "${USE_SHUNT_TCP}" = "1" ] && $ip6t_m -A PSW $(comment "默认") -p tcp $TCP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT6) -j DROP" [ "${TCP_PROXY_MODE}" != "disable" ] && $ip6t_m -A PSW $(comment "默认") -p tcp $TCP_PROXY_DROP_PORTS "-j DROP" } add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP" [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")" [ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS $(dst $IPSET_SHUNT) "-j DROP" [ "${TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "-j DROP" echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]" } [ "$UDP_PROXY_DROP_PORTS" != "disable" ] && { [ "$PROXY_IPV6" == "1" ] && { add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-d $FAKE_IP_6 -j DROP" [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK6) -j DROP" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_GFW6) -j DROP" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j DROP")" [ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT6) -j DROP" [ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-j DROP" } add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP" [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")" [ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT) -j DROP" [ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-j DROP" echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]" } } # 加载TCP默认代理模式 if [ -n "${TCP_PROXY_MODE}" ]; then [ -n "$TCP_NODE" ] && { msg2="${msg}使用 TCP 节点[$(config_n_get $TCP_NODE remarks)]" if [ -n "${is_tproxy}" ]; then msg2="${msg2}(TPROXY:${TCP_REDIR_PORT})" ipt_j="-j PSW_RULE" else msg2="${msg2}(REDIRECT:${TCP_REDIR_PORT})" ipt_j="$(REDIRECT $TCP_REDIR_PORT)" fi [ "$accept_icmp" = "1" ] && { $ipt_n -A PSW $(comment "默认") -p icmp -d $FAKE_IP $(REDIRECT) [ "${USE_PROXY_LIST}" = "1" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_BLACK) $(REDIRECT) [ "${USE_GFW_LIST}" = "1" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_GFW) $(REDIRECT) [ "${CHN_LIST}" != "0" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST}) [ "${USE_SHUNT_TCP}" = "1" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_SHUNT) $(REDIRECT) [ "${TCP_PROXY_MODE}" != "disable" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(REDIRECT) } [ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && { $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp -d $FAKE_IP_6 $(REDIRECT) [ "${USE_PROXY_LIST}" = "1" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_BLACK6) $(REDIRECT) [ "${USE_GFW_LIST}" = "1" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_GFW6) $(REDIRECT) [ "${CHN_LIST}" != "0" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST}) [ "${USE_SHUNT_TCP}" = "1" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_SHUNT6) $(REDIRECT) [ "${TCP_PROXY_MODE}" != "disable" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(REDIRECT) } $ipt_tmp -A PSW $(comment "默认") -p tcp -d $FAKE_IP ${ipt_j} [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK) ${ipt_j}" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW) ${ipt_j}" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "${ipt_j}")" [ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT) ${ipt_j}" [ "${TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "${ipt_j}" [ -n "${is_tproxy}" ]&& $ipt_tmp -A PSW $(comment "默认") -p tcp $(REDIRECT $TCP_REDIR_PORT TPROXY) [ "$PROXY_IPV6" == "1" ] && { $ip6t_m -A PSW $(comment "默认") -p tcp -d $FAKE_IP_6 -j PSW_RULE [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")" [ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE" [ "${TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "-j PSW_RULE" $ip6t_m -A PSW $(comment "默认") -p tcp $(REDIRECT $TCP_REDIR_PORT TPROXY) } echolog " - ${msg2}" } fi $ipt_n -A PSW $(comment "默认") -p tcp -j RETURN $ipt_m -A PSW $(comment "默认") -p tcp -j RETURN $ip6t_m -A PSW $(comment "默认") -p tcp -j RETURN # 加载UDP默认代理模式 if [ -n "${UDP_PROXY_MODE}" ]; then [ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ] && { msg2="${msg}使用 UDP 节点[$(config_n_get $UDP_NODE remarks)](TPROXY:${UDP_REDIR_PORT})" $ipt_m -A PSW $(comment "默认") -p udp -d $FAKE_IP -j PSW_RULE [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK) -j PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW) -j PSW_RULE" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")" [ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT) -j PSW_RULE" [ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "-j PSW_RULE" $ipt_m -A PSW $(comment "默认") -p udp $(REDIRECT $UDP_REDIR_PORT TPROXY) [ "$PROXY_IPV6" == "1" ] && { $ip6t_m -A PSW $(comment "默认") -p udp -d $FAKE_IP_6 -j PSW_RULE [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")" [ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE" [ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "-j PSW_RULE" $ip6t_m -A PSW $(comment "默认") -p udp $(REDIRECT $UDP_REDIR_PORT TPROXY) } echolog " - ${msg2}" } fi $ipt_m -A PSW $(comment "默认") -p udp -j RETURN $ip6t_m -A PSW $(comment "默认") -p udp -j RETURN } } filter_haproxy() { for item in ${haproxy_items}; do local ip=$(get_host_ip ipv4 $(echo $item | awk -F ":" '{print $1}') 1) ipset -q add $IPSET_VPS $ip done echolog " - [$?]加入负载均衡的节点到ipset[$IPSET_VPS]直连完成" } filter_vpsip() { uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([0-9]{1,3}[\.]){3}[0-9]{1,3}" | grep -v "^127\.0\.0\.1$" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_VPS &/g" | awk '{print $0} END{print "COMMIT"}' | ipset -! -R echolog " - [$?]加入所有IPv4节点到ipset[$IPSET_VPS]直连完成" uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_VPS6 &/g" | awk '{print $0} END{print "COMMIT"}' | ipset -! -R echolog " - [$?]加入所有IPv6节点到ipset[$IPSET_VPS6]直连完成" } filter_server_port() { local address=${1} local port=${2} local stream=${3} stream=$(echo ${3} | tr 'A-Z' 'a-z') local _is_tproxy ipt_tmp ipt_tmp=$ipt_n _is_tproxy=${is_tproxy} [ "$stream" == "udp" ] && _is_tproxy="TPROXY" [ -n "${_is_tproxy}" ] && ipt_tmp=$ipt_m for _ipt in 4 6; do [ "$_ipt" == "4" ] && _ipt=$ipt_tmp [ "$_ipt" == "6" ] && _ipt=$ip6t_m $_ipt -n -L PSW_OUTPUT | grep -q "${address}:${port}" if [ $? -ne 0 ]; then $_ipt -I PSW_OUTPUT $(comment "${address}:${port}") -p $stream -d $address --dport $port -j RETURN 2>/dev/null fi done } filter_node() { local node=${1} local stream=${2} if [ -n "$node" ]; then local address=$(config_n_get $node address) local port=$(config_n_get $node port) [ -z "$address" ] && [ -z "$port" ] && { return 1 } filter_server_port $address $port $stream filter_server_port $address $port $stream fi } filter_direct_node_list() { [ ! -s "$TMP_PATH/direct_node_list" ] && return for _node_id in $(cat $TMP_PATH/direct_node_list | awk '!seen[$0]++'); do filter_node "$_node_id" TCP filter_node "$_node_id" UDP unset _node_id done } add_firewall_rule() { echolog "开始加载防火墙规则..." ipset -! create $IPSET_LOCAL nethash maxelem 1048576 ipset -! create $IPSET_LAN nethash maxelem 1048576 ipset -! create $IPSET_VPS nethash maxelem 1048576 ipset -! create $IPSET_SHUNT nethash maxelem 1048576 timeout 172800 ipset -! create $IPSET_GFW nethash maxelem 1048576 timeout 172800 ipset -! create $IPSET_CHN nethash maxelem 1048576 timeout 172800 ipset -! create $IPSET_BLACK nethash maxelem 1048576 timeout 172800 ipset -! create $IPSET_WHITE nethash maxelem 1048576 timeout 172800 ipset -! create $IPSET_BLOCK nethash maxelem 1048576 timeout 172800 ipset -! create $IPSET_LOCAL6 nethash family inet6 maxelem 1048576 ipset -! create $IPSET_LAN6 nethash family inet6 maxelem 1048576 ipset -! create $IPSET_VPS6 nethash family inet6 maxelem 1048576 ipset -! create $IPSET_SHUNT6 nethash family inet6 maxelem 1048576 timeout 172800 ipset -! create $IPSET_GFW6 nethash family inet6 maxelem 1048576 timeout 172800 ipset -! create $IPSET_CHN6 nethash family inet6 maxelem 1048576 timeout 172800 ipset -! create $IPSET_BLACK6 nethash family inet6 maxelem 1048576 timeout 172800 ipset -! create $IPSET_WHITE6 nethash family inet6 maxelem 1048576 timeout 172800 ipset -! create $IPSET_BLOCK6 nethash family inet6 maxelem 1048576 timeout 172800 cat $RULES_PATH/chnroute | tr -s '\n' | grep -v "^#" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_CHN &/g" -e "s/$/ timeout 0/g" | ipset -! -R cat $RULES_PATH/chnroute6 | tr -s '\n' | grep -v "^#" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_CHN6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R #导入规则列表、分流规则中的IP列表 local USE_SHUNT_NODE=0 local USE_PROXY_LIST_ALL=${USE_PROXY_LIST} local USE_DIRECT_LIST_ALL=${USE_DIRECT_LIST} local USE_BLOCK_LIST_ALL=${USE_BLOCK_LIST} local _TCP_NODE=$(config_t_get global tcp_node) local _UDP_NODE=$(config_t_get global udp_node) local USE_GEOVIEW=$(config_t_get global_rules enable_geoview) [ -n "$_TCP_NODE" ] && [ "$(config_n_get $_TCP_NODE protocol)" = "_shunt" ] && USE_SHUNT_TCP=1 && USE_SHUNT_NODE=1 [ -n "$_UDP_NODE" ] && [ "$(config_n_get $_UDP_NODE protocol)" = "_shunt" ] && USE_SHUNT_UDP=1 && USE_SHUNT_NODE=1 [ "$_UDP_NODE" = "tcp" ] && USE_SHUNT_UDP=$USE_SHUNT_TCP for acl_section in $(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do [ "$(config_n_get $acl_section enabled)" != "1" ] && continue [ "$(config_n_get $acl_section use_global_config 0)" != "1" ] && { [ "$(config_n_get $acl_section use_direct_list 1)" = "1" ] && USE_PROXY_LIST_ALL=1 [ "$(config_n_get $acl_section use_proxy_list 1)" = "1" ] && USE_DIRECT_LIST_ALL=1 [ "$(config_n_get $acl_section use_block_list 1)" = "1" ] && USE_BLOCK_LIST_ALL=1 } for _node in $(config_n_get $acl_section tcp_node) $(config_n_get $acl_section udp_node); do local node_protocol=$(config_n_get $_node protocol) [ "$node_protocol" = "_shunt" ] && { USE_SHUNT_NODE=1; break; } done done #直连列表 [ "$USE_DIRECT_LIST_ALL" = "1" ] && { cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_WHITE &/g" -e "s/$/ timeout 0/g" | ipset -! -R cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_WHITE6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R [ "$USE_GEOVIEW" = "1" ] && { local GEOIP_CODE=$(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_WHITE &/g" -e "s/$/ timeout 0/g" | ipset -! -R get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_WHITE6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R echolog " - [$?]解析并加入[直连列表] GeoIP 到 IPSET 完成" fi } } #代理列表 [ "$USE_PROXY_LIST_ALL" = "1" ] && { cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLACK &/g" -e "s/$/ timeout 0/g" | ipset -! -R cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLACK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R [ "$USE_GEOVIEW" = "1" ] && { local GEOIP_CODE=$(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLACK &/g" -e "s/$/ timeout 0/g" | ipset -! -R get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLACK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R echolog " - [$?]解析并加入[代理列表] GeoIP 到 IPSET 完成" fi } } #屏蔽列表 [ "$USE_BLOCK_LIST_ALL" = "1" ] && { cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLOCK &/g" -e "s/$/ timeout 0/g" | ipset -! -R cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLOCK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R [ "$USE_GEOVIEW" = "1" ] && { local GEOIP_CODE=$(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLOCK &/g" -e "s/$/ timeout 0/g" | ipset -! -R get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLOCK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R echolog " - [$?]解析并加入[屏蔽列表] GeoIP 到 IPSET 完成" fi } } #分流列表 [ "$USE_SHUNT_NODE" = "1" ] && { local GEOIP_CODE="" local shunt_ids=$(uci show $CONFIG | grep "=shunt_rules" | awk -F '.' '{print $2}' | awk -F '=' '{print $1}') for shunt_id in $shunt_ids; do config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_SHUNT &/g" -e "s/$/ timeout 0/g" | ipset -! -R config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_SHUNT6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R [ "$USE_GEOVIEW" = "1" ] && { local geoip_code=$(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g') [ -n "$geoip_code" ] && GEOIP_CODE="${GEOIP_CODE:+$GEOIP_CODE,}$geoip_code" } done if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_SHUNT &/g" -e "s/$/ timeout 0/g" | ipset -! -R get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_SHUNT6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R echolog " - [$?]解析并加入[分流节点] GeoIP 到 IPSET 完成" fi } ipset -! -R <<-EOF $(ip address show | grep -w "inet" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/^/add $IPSET_LOCAL /") EOF ipset -! -R <<-EOF $(ip address show | grep -w "inet6" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/^/add $IPSET_LOCAL6 /") EOF #局域网IP列表 ipset -! -R <<-EOF $(gen_lanlist | sed -e "s/^/add $IPSET_LAN /") EOF ipset -! -R <<-EOF $(gen_lanlist_6 | sed -e "s/^/add $IPSET_LAN6 /") EOF # 忽略特殊IP段 local lan_ifname lan_ip lan_ifname=$(uci -q -p /tmp/state get network.lan.ifname) [ -n "$lan_ifname" ] && { lan_ip=$(ip address show $lan_ifname | grep -w "inet" | awk '{print $2}') lan_ip6=$(ip address show $lan_ifname | grep -w "inet6" | awk '{print $2}') #echolog "本机IPv4网段互访直连:${lan_ip}" #echolog "本机IPv6网段互访直连:${lan_ip6}" [ -n "$lan_ip" ] && ipset -! -R <<-EOF $(echo $lan_ip | sed -e "s/ /\n/g" | sed -e "s/^/add $IPSET_LAN /") EOF [ -n "$lan_ip6" ] && ipset -! -R <<-EOF $(echo $lan_ip6 | sed -e "s/ /\n/g" | sed -e "s/^/add $IPSET_LAN6 /") EOF } [ -n "$ISP_DNS" ] && { #echolog "处理 ISP DNS 例外..." for ispip in $ISP_DNS; do ipset -! add $IPSET_WHITE $ispip timeout 0 echolog " - [$?]追加ISP IPv4 DNS到白名单:${ispip}" done } [ -n "$ISP_DNS6" ] && { #echolog "处理 ISP IPv6 DNS 例外..." for ispip6 in $ISP_DNS6; do ipset -! add $IPSET_WHITE6 $ispip6 timeout 0 echolog " - [$?]追加ISP IPv6 DNS到白名单:${ispip6}" done } # 过滤所有节点IP filter_vpsip > /dev/null 2>&1 & # filter_haproxy > /dev/null 2>&1 & accept_icmp=$(config_t_get global_forwarding accept_icmp 0) accept_icmpv6=$(config_t_get global_forwarding accept_icmpv6 0) if [ "${TCP_PROXY_WAY}" = "redirect" ]; then unset is_tproxy elif [ "${TCP_PROXY_WAY}" = "tproxy" ]; then is_tproxy="TPROXY" fi $ipt_n -N PSW $ipt_n -A PSW $(dst $IPSET_LAN) -j RETURN $ipt_n -A PSW $(dst $IPSET_VPS) -j RETURN WAN_IP=$(get_wan_ip) [ ! -z "${WAN_IP}" ] && $ipt_n -A PSW $(comment "WAN_IP_RETURN") -d "${WAN_IP}" -j RETURN [ "$accept_icmp" = "1" ] && insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p icmp -j PSW" [ -z "${is_tproxy}" ] && insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p tcp -j PSW" $ipt_n -N PSW_OUTPUT $ipt_n -A PSW_OUTPUT $(dst $IPSET_LAN) -j RETURN $ipt_n -A PSW_OUTPUT $(dst $IPSET_VPS) -j RETURN [ "${USE_DIRECT_LIST}" = "1" ] && $ipt_n -A PSW_OUTPUT $(dst $IPSET_WHITE) -j RETURN $ipt_n -A PSW_OUTPUT -m mark --mark 0xff -j RETURN $ipt_n -N PSW_DNS if [ $(config_t_get global dns_redirect "1") = "0" ]; then #Only hijack when dest address is local IP $ipt_n -I PREROUTING $(dst $IPSET_LOCAL) -j PSW_DNS else $ipt_n -I PREROUTING 1 -j PSW_DNS fi $ipt_m -N PSW_DIVERT $ipt_m -A PSW_DIVERT -j MARK --set-mark 1 $ipt_m -A PSW_DIVERT -j ACCEPT $ipt_m -N PSW_RULE $ipt_m -A PSW_RULE -j CONNMARK --restore-mark $ipt_m -A PSW_RULE -m mark --mark 1 -j RETURN $ipt_m -A PSW_RULE -p tcp -m tcp --tcp-flags FIN,SYN,RST,ACK SYN -j MARK --set-xmark 1 $ipt_m -A PSW_RULE -p udp -m conntrack --ctstate NEW -j MARK --set-xmark 1 $ipt_m -A PSW_RULE -j CONNMARK --save-mark $ipt_m -N PSW $ipt_m -A PSW $(dst $IPSET_LAN) -j RETURN $ipt_m -A PSW $(dst $IPSET_VPS) -j RETURN [ ! -z "${WAN_IP}" ] && { $ipt_m -A PSW $(comment "WAN_IP_RETURN") -d "${WAN_IP}" -j RETURN echolog " - [$?]追加WAN IP到iptables:${WAN_IP}" } unset WAN_IP insert_rule_before "$ipt_m" "PREROUTING" "mwan3" "-j PSW" insert_rule_before "$ipt_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT" $ipt_m -N PSW_OUTPUT $ipt_m -A PSW_OUTPUT $(dst $IPSET_LAN) -j RETURN $ipt_m -A PSW_OUTPUT $(dst $IPSET_VPS) -j RETURN [ -n "$IPT_APPEND_DNS" ] && { local local_dns dns_address dns_port for local_dns in $(echo $IPT_APPEND_DNS | tr ',' ' '); do dns_address=$(echo "$local_dns" | sed -E 's/(@|\[)?([0-9a-fA-F:.]+)(@|#|$).*/\2/') dns_port=$(echo "$local_dns" | sed -nE 's/.*#([0-9]+)$/\1/p') if echo "$dns_address" | grep -q -v ':'; then $ipt_m -A PSW_OUTPUT -p udp -d ${dns_address} --dport ${dns_port:-53} -j RETURN $ipt_m -A PSW_OUTPUT -p tcp -d ${dns_address} --dport ${dns_port:-53} -j RETURN echolog " - [$?]追加直连DNS到iptables:${dns_address}:${dns_port:-53}" else $ip6t_m -A PSW_OUTPUT -p udp -d ${dns_address} --dport ${dns_port:-53} -j RETURN $ip6t_m -A PSW_OUTPUT -p tcp -d ${dns_address} --dport ${dns_port:-53} -j RETURN echolog " - [$?]追加直连DNS到iptables:[${dns_address}]:${dns_port:-53}" fi done } [ "${USE_BLOCK_LIST}" = "1" ] && $ipt_m -A PSW_OUTPUT $(dst $IPSET_BLOCK) -j DROP [ "${USE_DIRECT_LIST}" = "1" ] && $ipt_m -A PSW_OUTPUT $(dst $IPSET_WHITE) -j RETURN $ipt_m -A PSW_OUTPUT -m mark --mark 0xff -j RETURN ip rule add fwmark 1 lookup 100 ip route add local 0.0.0.0/0 dev lo table 100 [ "$accept_icmpv6" = "1" ] && { $ip6t_n -N PSW $ip6t_n -A PSW $(dst $IPSET_LAN6) -j RETURN $ip6t_n -A PSW $(dst $IPSET_VPS6) -j RETURN $ip6t_n -A PREROUTING -p ipv6-icmp -j PSW $ip6t_n -N PSW_OUTPUT $ip6t_n -A PSW_OUTPUT $(dst $IPSET_LAN6) -j RETURN $ip6t_n -A PSW_OUTPUT $(dst $IPSET_VPS6) -j RETURN [ "${USE_DIRECT_LIST}" = "1" ] && $ip6t_n -A PSW_OUTPUT $(dst $IPSET_WHITE6) -j RETURN $ip6t_n -A PSW_OUTPUT -m mark --mark 0xff -j RETURN } $ip6t_n -N PSW_DNS if [ $(config_t_get global dns_redirect "1") = "0" ]; then #Only hijack when dest address is local IP $ip6t_n -I PREROUTING $(dst $IPSET_LOCAL6) -j PSW_DNS else $ip6t_n -I PREROUTING 1 -j PSW_DNS fi $ip6t_m -N PSW_DIVERT $ip6t_m -A PSW_DIVERT -j MARK --set-mark 1 $ip6t_m -A PSW_DIVERT -j ACCEPT $ip6t_m -N PSW_RULE $ip6t_m -A PSW_RULE -j CONNMARK --restore-mark $ip6t_m -A PSW_RULE -m mark --mark 1 -j RETURN $ip6t_m -A PSW_RULE -p tcp -m tcp --tcp-flags FIN,SYN,RST,ACK SYN -j MARK --set-xmark 1 $ip6t_m -A PSW_RULE -p udp -m conntrack --ctstate NEW -j MARK --set-xmark 1 $ip6t_m -A PSW_RULE -j CONNMARK --save-mark $ip6t_m -N PSW $ip6t_m -A PSW $(dst $IPSET_LAN6) -j RETURN $ip6t_m -A PSW $(dst $IPSET_VPS6) -j RETURN WAN6_IP=$(get_wan6_ip) [ ! -z "${WAN6_IP}" ] && $ip6t_m -A PSW $(comment "WAN6_IP_RETURN") -d ${WAN6_IP} -j RETURN unset WAN6_IP insert_rule_before "$ip6t_m" "PREROUTING" "mwan3" "-j PSW" insert_rule_before "$ip6t_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT" $ip6t_m -N PSW_OUTPUT $ip6t_m -A PSW_OUTPUT -m mark --mark 0xff -j RETURN $ip6t_m -A PSW_OUTPUT $(dst $IPSET_LAN6) -j RETURN $ip6t_m -A PSW_OUTPUT $(dst $IPSET_VPS6) -j RETURN [ "${USE_BLOCK_LIST}" = "1" ] && $ip6t_m -A PSW_OUTPUT $(dst $IPSET_BLOCK6) -j DROP [ "${USE_DIRECT_LIST}" = "1" ] && $ip6t_m -A PSW_OUTPUT $(dst $IPSET_WHITE6) -j RETURN ip -6 rule add fwmark 1 table 100 ip -6 route add local ::/0 dev lo table 100 [ "$TCP_UDP" = "1" ] && [ -z "$UDP_NODE" ] && UDP_NODE=$TCP_NODE [ "$ENABLED_DEFAULT_ACL" == 1 ] && { local ipt_tmp=$ipt_n if [ -n "${is_tproxy}" ]; then ipt_tmp=$ipt_m ipt_j="-j PSW_RULE" else ipt_j="$(REDIRECT $TCP_REDIR_PORT)" fi msg="【路由器本机】," [ "$TCP_NO_REDIR_PORTS" != "disable" ] && { add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN" add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN" if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]" else unset LOCALHOST_TCP_PROXY_MODE echolog " - ${msg}不代理所有 TCP 端口" fi } [ "$UDP_NO_REDIR_PORTS" != "disable" ] && { add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_NO_REDIR_PORTS "-j RETURN" add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_NO_REDIR_PORTS "-j RETURN" if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]" else unset LOCALHOST_UDP_PROXY_MODE echolog " - ${msg}不代理所有 UDP 端口" fi } if ([ -n "$TCP_NODE" ] && [ -n "${LOCALHOST_TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${LOCALHOST_UDP_PROXY_MODE}" ]); then [ -n "$DNS_REDIRECT_PORT" ] && { $ipt_n -A OUTPUT $(comment "PSW_DNS") -p udp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT $ip6t_n -A OUTPUT $(comment "PSW_DNS") -p udp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT 2>/dev/null $ipt_n -A OUTPUT $(comment "PSW_DNS") -p tcp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT $ip6t_n -A OUTPUT $(comment "PSW_DNS") -p tcp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT 2>/dev/null } fi [ -n "${LOCALHOST_TCP_PROXY_MODE}" -o -n "${LOCALHOST_UDP_PROXY_MODE}" ] && { [ "$TCP_PROXY_DROP_PORTS" != "disable" ] && { add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP" [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")" [ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT) -j DROP" [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "-j DROP" echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]" } [ "$UDP_PROXY_DROP_PORTS" != "disable" ] && { add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP" [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")" [ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT) -j DROP" [ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "-j DROP" echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]" } } # 加载路由器自身代理 TCP if [ -n "$TCP_NODE" ]; then _proxy_tcp_access() { [ -n "${2}" ] || return 0 if echo "${2}" | grep -q -v ':'; then ipset -q test $IPSET_LAN ${2} [ $? -eq 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问" return 0 } if [ -z "${is_tproxy}" ]; then $ipt_n -I PSW_OUTPUT -p tcp -d ${2} --dport ${3} $(REDIRECT $TCP_REDIR_PORT) else $ipt_m -I PSW_OUTPUT -p tcp -d ${2} --dport ${3} -j PSW_RULE $ipt_m -I PSW $(comment "本机") -p tcp -i lo -d ${2} --dport ${3} $(REDIRECT $TCP_REDIR_PORT TPROXY) fi echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 TCP 转发链" else ipset -q test $IPSET_LAN6 ${2} [ $? -eq 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问" return 0 } $ip6t_m -I PSW_OUTPUT -p tcp -d ${2} --dport ${3} -j PSW_RULE $ip6t_m -I PSW $(comment "本机") -p tcp -i lo -d ${2} --dport ${3} $(REDIRECT $TCP_REDIR_PORT TPROXY) echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 TCP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!" fi } [ -n "${TCP_PROXY_DNS}" ] && hosts_foreach REMOTE_DNS _proxy_tcp_access 53 [ "$accept_icmp" = "1" ] && { $ipt_n -A OUTPUT -p icmp -j PSW_OUTPUT $ipt_n -A PSW_OUTPUT -p icmp -d $FAKE_IP $(REDIRECT) [ "${USE_PROXY_LIST}" = "1" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_BLACK) $(REDIRECT) [ "${USE_GFW_LIST}" = "1" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_GFW) $(REDIRECT) [ "${CHN_LIST}" != "0" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST}) [ "${USE_SHUNT_TCP}" = "1" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_SHUNT) $(REDIRECT) [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && $ipt_n -A PSW_OUTPUT -p icmp $(REDIRECT) } [ "$accept_icmpv6" = "1" ] && { $ip6t_n -A OUTPUT -p ipv6-icmp -j PSW_OUTPUT $ip6t_n -A PSW_OUTPUT -p ipv6-icmp -d $FAKE_IP_6 $(REDIRECT) [ "${USE_PROXY_LIST}" = "1" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_BLACK6) $(REDIRECT) [ "${USE_GFW_LIST}" = "1" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_GFW6) $(REDIRECT) [ "${CHN_LIST}" != "0" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST}) [ "${USE_SHUNT_TCP}" = "1" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_SHUNT6) $(REDIRECT) [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(REDIRECT) } [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && { $ipt_tmp -A PSW_OUTPUT -p tcp -d $FAKE_IP ${ipt_j} [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK) ${ipt_j}" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW) ${ipt_j}" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "${ipt_j}")" [ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT) ${ipt_j}" [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "${ipt_j}" [ -n "${is_tproxy}" ] && $ipt_m -A PSW $(comment "本机") -p tcp -i lo $(REDIRECT $TCP_REDIR_PORT TPROXY) } [ -z "${is_tproxy}" ] && $ipt_n -A OUTPUT -p tcp -j PSW_OUTPUT [ -n "${is_tproxy}" ] && { $ipt_m -A PSW $(comment "本机") -p tcp -i lo -j RETURN insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p tcp -j PSW_OUTPUT" } [ "$PROXY_IPV6" == "1" ] && { [ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && { $ip6t_m -A PSW_OUTPUT -p tcp -d $FAKE_IP_6 -j PSW_RULE [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")" [ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE" [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "-j PSW_RULE" $ip6t_m -A PSW $(comment "本机") -p tcp -i lo $(REDIRECT $TCP_REDIR_PORT TPROXY) } $ip6t_m -A PSW $(comment "本机") -p tcp -i lo -j RETURN insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p tcp -j PSW_OUTPUT" } fi # 加载路由器自身代理 UDP if [ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ]; then _proxy_udp_access() { [ -n "${2}" ] || return 0 if echo "${2}" | grep -q -v ':'; then ipset -q test $IPSET_LAN ${2} [ $? == 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问" return 0 } $ipt_m -I PSW_OUTPUT -p udp -d ${2} --dport ${3} -j PSW_RULE $ipt_m -I PSW $(comment "本机") -p udp -i lo -d ${2} --dport ${3} $(REDIRECT $UDP_REDIR_PORT TPROXY) echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 UDP 转发链" else ipset -q test $IPSET_LAN6 ${2} [ $? == 0 ] && { echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问" return 0 } $ip6t_m -I PSW_OUTPUT -p udp -d ${2} --dport ${3} -j PSW_RULE $ip6t_m -I PSW $(comment "本机") -p udp -i lo -d ${2} --dport ${3} $(REDIRECT $UDP_REDIR_PORT TPROXY) echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 UDP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!" fi } [ -n "${UDP_PROXY_DNS}" ] && hosts_foreach REMOTE_DNS _proxy_udp_access 53 [ -n "${LOCALHOST_UDP_PROXY_MODE}" ] && { $ipt_m -A PSW_OUTPUT -p udp -d $FAKE_IP -j PSW_RULE [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK) -j PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW) -j PSW_RULE" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")" [ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT) -j PSW_RULE" [ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "-j PSW_RULE" $ipt_m -A PSW $(comment "本机") -p udp -i lo $(REDIRECT $UDP_REDIR_PORT TPROXY) } $ipt_m -A PSW $(comment "本机") -p udp -i lo -j RETURN insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p udp -j PSW_OUTPUT" [ "$PROXY_IPV6" == "1" ] && { [ -n "$LOCALHOST_UDP_PROXY_MODE" ] && { $ip6t_m -A PSW_OUTPUT -p udp -d $FAKE_IP_6 -j PSW_RULE [ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE" [ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE" [ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")" [ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE" [ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "-j PSW_RULE" $ip6t_m -A PSW $(comment "本机") -p udp -i lo $(REDIRECT $UDP_REDIR_PORT TPROXY) } $ip6t_m -A PSW $(comment "本机") -p udp -i lo -j RETURN insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p udp -j PSW_OUTPUT" } fi $ipt_m -I OUTPUT $(comment "mangle-OUTPUT-PSW") -o lo -j RETURN insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -m mark --mark 1 -j RETURN" $ip6t_m -I OUTPUT $(comment "mangle-OUTPUT-PSW") -o lo -j RETURN insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -m mark --mark 1 -j RETURN" } # 加载ACLS load_acl [ -d "${TMP_IFACE_PATH}" ] && { for iface in $(ls ${TMP_IFACE_PATH}); do $ipt_n -I PSW_OUTPUT -o $iface -j RETURN $ipt_m -I PSW_OUTPUT -o $iface -j RETURN done } $ipt_n -I PREROUTING $(comment "PSW") -m mark --mark 1 -j RETURN $ip6t_n -I PREROUTING $(comment "PSW") -m mark --mark 1 -j RETURN filter_direct_node_list > /dev/null 2>&1 & echolog "防火墙规则加载完成!" } del_firewall_rule() { for ipt in "$ipt_n" "$ipt_m" "$ip6t_n" "$ip6t_m"; do for chain in "PREROUTING" "OUTPUT"; do for i in $(seq 1 $($ipt -nL $chain | grep -c PSW)); do local index=$($ipt --line-number -nL $chain | grep PSW | head -1 | awk '{print $1}') $ipt -D $chain $index 2>/dev/null done done for chain in "PSW" "PSW_OUTPUT" "PSW_DIVERT" "PSW_DNS" "PSW_RULE"; do $ipt -F $chain 2>/dev/null $ipt -X $chain 2>/dev/null done done ip rule del fwmark 1 lookup 100 2>/dev/null ip route del local 0.0.0.0/0 dev lo table 100 2>/dev/null ip -6 rule del fwmark 1 table 100 2>/dev/null ip -6 route del local ::/0 dev lo table 100 2>/dev/null destroy_ipset $IPSET_LOCAL destroy_ipset $IPSET_LAN destroy_ipset $IPSET_VPS destroy_ipset $IPSET_SHUNT #destroy_ipset $IPSET_GFW #destroy_ipset $IPSET_CHN #destroy_ipset $IPSET_BLACK destroy_ipset $IPSET_BLOCK destroy_ipset $IPSET_WHITE destroy_ipset $IPSET_LOCAL6 destroy_ipset $IPSET_LAN6 destroy_ipset $IPSET_VPS6 destroy_ipset $IPSET_SHUNT6 #destroy_ipset $IPSET_GFW6 #destroy_ipset $IPSET_CHN6 #destroy_ipset $IPSET_BLACK6 destroy_ipset $IPSET_BLOCK6 destroy_ipset $IPSET_WHITE6 $DIR/app.sh echolog "删除 iptables 规则完成。" } flush_ipset() { $DIR/app.sh echolog "清空 IPSet。" for _name in $(ipset list | grep "Name: " | grep "passwall_" | awk '{print $2}'); do destroy_ipset ${_name} done } flush_include() { echo '#!/bin/sh' >$FWI } gen_include() { flush_include extract_rules() { local _ipt="${ipt}" [ "$1" == "6" ] && _ipt="${ip6t}" [ -z "${_ipt}" ] && return echo "*$2" ${_ipt}-save -t $2 | grep "PSW" | grep -v "\-j PSW$" | grep -v "mangle\-OUTPUT\-PSW" | grep -v "socket \-j PSW_DIVERT$" | sed -e "s/^-A \(OUTPUT\|PREROUTING\)/-I \1 1/" echo 'COMMIT' } local __ipt="" [ -n "${ipt}" ] && { __ipt=$(cat <<- EOF mangle_output_psw=\$(${ipt}-save -t mangle | grep "PSW" | grep "mangle\-OUTPUT\-PSW" | sed "s#-A OUTPUT ##g") $ipt-save -c | grep -v "PSW" | $ipt-restore -c $ipt-restore -n <<-EOT $(extract_rules 4 nat) $(extract_rules 4 mangle) EOT echo "\${mangle_output_psw}" | while read line; do \$(${MY_PATH} insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "\${line}") done [ "$accept_icmp" = "1" ] && \$(${MY_PATH} insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p icmp -j PSW") [ -z "${is_tproxy}" ] && \$(${MY_PATH} insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p tcp -j PSW") \$(${MY_PATH} insert_rule_before "$ipt_m" "PREROUTING" "mwan3" "-j PSW") \$(${MY_PATH} insert_rule_before "$ipt_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT") WAN_IP=\$(${MY_PATH} get_wan_ip) PR_INDEX=\$(${MY_PATH} RULE_LAST_INDEX "$ipt_n" PSW WAN_IP_RETURN -1) if [ \$PR_INDEX -ge 0 ]; then [ ! -z "\${WAN_IP}" ] && $ipt_n -R PSW \$PR_INDEX $(comment "WAN_IP_RETURN") -d "\${WAN_IP}" -j RETURN fi PR_INDEX=\$(${MY_PATH} RULE_LAST_INDEX "$ipt_m" PSW WAN_IP_RETURN -1) if [ \$PR_INDEX -ge 0 ]; then [ ! -z "\${WAN_IP}" ] && $ipt_m -R PSW \$PR_INDEX $(comment "WAN_IP_RETURN") -d "\${WAN_IP}" -j RETURN fi EOF ) } local __ip6t="" [ -n "${ip6t}" ] && { __ip6t=$(cat <<- EOF mangle_output_psw=\$(${ip6t}-save -t mangle | grep "PSW" | grep "mangle\-OUTPUT\-PSW" | sed "s#-A OUTPUT ##g") $ip6t-save -c | grep -v "PSW" | $ip6t-restore -c $ip6t-restore -n <<-EOT $(extract_rules 6 nat) $(extract_rules 6 mangle) EOT echo "\${mangle_output_psw}" | while read line; do \$(${MY_PATH} insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "\${line}") done [ "$accept_icmpv6" = "1" ] && $ip6t_n -A PREROUTING -p ipv6-icmp -j PSW \$(${MY_PATH} insert_rule_before "$ip6t_m" "PREROUTING" "mwan3" "-j PSW") \$(${MY_PATH} insert_rule_before "$ip6t_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT") PR_INDEX=\$(${MY_PATH} RULE_LAST_INDEX "$ip6t_m" PSW WAN6_IP_RETURN -1) if [ \$PR_INDEX -ge 0 ]; then WAN6_IP=\$(${MY_PATH} get_wan6_ip) [ ! -z "\${WAN6_IP}" ] && $ip6t_m -R PSW \$PR_INDEX $(comment "WAN6_IP_RETURN") -d "\${WAN6_IP}" -j RETURN fi EOF ) } cat <<-EOF >> $FWI ${__ipt} ${__ip6t} return 0 EOF return 0 } get_ipt_bin() { echo $ipt } get_ip6t_bin() { echo $ip6t } start() { [ "$ENABLED_DEFAULT_ACL" == 0 -a "$ENABLED_ACLS" == 0 ] && return add_firewall_rule gen_include } stop() { del_firewall_rule [ $(config_t_get global flush_set "0") = "1" ] && { uci -q delete ${CONFIG}.@global[0].flush_set uci -q commit ${CONFIG} flush_ipset rm -rf /tmp/etc/passwall_tmp/singbox* rm -rf /tmp/etc/passwall_tmp/dnsmasq* } flush_include } arg1=$1 shift case $arg1 in RULE_LAST_INDEX) RULE_LAST_INDEX "$@" ;; insert_rule_before) insert_rule_before "$@" ;; insert_rule_after) insert_rule_after "$@" ;; get_ipt_bin) get_ipt_bin ;; get_ip6t_bin) get_ip6t_bin ;; get_wan_ip) get_wan_ip ;; get_wan6_ip) get_wan6_ip ;; filter_direct_node_list) filter_direct_node_list ;; stop) stop ;; start) start ;; *) ;; esac
294coder/Efficient-MIF
6,593
utils/save_checker.py
# author: Zihan # date: 2023/11/12 from dataclasses import dataclass import numpy as np from pathlib import Path import shutil from functools import partial import sys sys.path.append('./') from utils.log_utils import easy_logger from utils.misc import is_none logger = easy_logger(func_name='save_checker') @dataclass() class BestMetricSaveChecker: _best_metric: "float | str" metric_name: "str | list[str]" check_order: str def __init__(self, metric_name: "str | list[str] | dict[str, float]"=None, check_order: str=None, *, avg_metrics_weights: dict[str, float]=None): # overide the avg_metrics_weights if it is None if is_none(avg_metrics_weights): if isinstance(metric_name, (list, tuple)): self.metric_name = metric_name avg_metrics_weights = {m: 1. for m in metric_name} elif isinstance(metric_name, dict): self.metric_name = list(metric_name.keys()) avg_metrics_weights = metric_name elif isinstance(metric_name, str): self.metric_name = metric_name avg_metrics_weights = {metric_name: 1.} else: raise ValueError(f'@metric_name should be a str or a list of str or a dict of str->float but got {metric_name}') else: self.metric_name = metric_name if is_none(metric_name): logger.warning(f'No metric name provided, and SaveChecker will alway return True') assert check_order in ['up', 'down', None] if check_order is None: check_order = self._default_setting() self.check_order = check_order self.avg_metrics_weights = avg_metrics_weights self.avg_metrics = isinstance(self.metric_name, list) and avg_metrics_weights is not None logger.info(f'SaveChecker initialized with metric_name={self.metric_name}, ', f'check_order={check_order}, avg_metrics_weights={avg_metrics_weights}') if isinstance(metric_name, list): assert not is_none(avg_metrics_weights), \ f'@avg_metrics_weights should not be None when @metric_name is a list' for m in metric_name: assert m in avg_metrics_weights.keys(), \ f'@avg_metrics_weights should have key {m} but got {avg_metrics_weights}' if check_order == 'up': default_best_metric_val = -np.Inf elif check_order == 'down': default_best_metric_val = np.Inf else: default_best_metric_val = 'none' self._best_metric = default_best_metric_val self._check_fn = (lambda new, old: new > old) if check_order=='up' else \ (lambda new, old: new <= old) def _default_setting(self): metric_name = self.metric_name.lower() if is_none(metric_name): return 'none' _default_dict = { 'psnr': 'up', 'ssim': 'up', 'sam': 'down', 'ergas': 'down', 'cc': 'up', 'scc': 'up', 'sd': 'up', 'en': 'up', 'sf': 'up', 'vif': 'up', 'mse': 'down', 'ag': 'up', 'mi': 'up', } if check_order := _default_dict.get(metric_name): if check_order is None: raise ValueError(f'No default setting for metric {metric_name}, ' + 'you should provide @check_order manually') else: return check_order def __call__(self, val_metrics: dict[str, float], *args): if is_none(self.metric_name): return True if self.avg_metrics: for m in self.metric_name: assert m in val_metrics.keys(), f'@val_metrics should have key {m} but got {val_metrics}' else: assert self.metric_name in val_metrics.keys(), f'@val_metrics should have key {self.metric_name} but got {val_metrics}' if not self.avg_metrics: new_val = val_metrics[self.metric_name] else: new_val = np.sum([val_metrics[k] * self.avg_metrics_weights[k] for k in self.metric_name]) prev_val = self._best_metric _save = self._check_fn(new_val, prev_val) if _save: self._best_metric = new_val return _save @property def best_metric(self): return self._best_metric # TODO: need test class SavedWeightsNumMonitor: def __init__(self, saved_weight_path: str, glob_pattern: str='ep*/', sort_saved_fn: callable=None, kept_num: int=4, rm_verbose: bool=False): self.saved_weight_path = Path(saved_weight_path) self.glob_pattern = glob_pattern self.sort_saved_fn = sort_saved_fn self.kept_num = kept_num self.rm_verbose = rm_verbose if rm_verbose: self.logger = easy_logger() @property def saved_weight_path(self): return self.saved_weight_path.glob(self.glob_pattern) @property def saved_num(self): # excluse .pth in the base dir return len(list(self.saved_weight_path)) def rm_for_keep_k(self): if self.saved_num > self.kept_num: saved_name = list(map(lambda x: x.name, self.saved_weight_path)) sort_fn = self.sort_saved_fn if self.sort_saved_fn is not None else \ partial(sorted, key=lambda x: int(x)) sorted_saved_name = sort_fn(saved_name) rm_saved_name = sorted_saved_name[self.kept_num:] for p in rm_saved_name: shutil.rmtree(self.saved_weight_path / p) if self.rm_verbose: self.logger.print(f'remove {p}') def __call__(self): self.rm_for_keep_k() if __name__ == '__main__': checker = BestMetricSaveChecker({"psnr": 0.1, "ssim": 0.2}, 'up') val_d1 = {'sam': 2.3, 'psnr': 10, 'ssim':0.8} val_d2 = {'sam': 2.4, 'psnr': 12, 'ssim':0.9} print(checker.best_metric) print(checker.metric_name) print(checker(val_d1)) print(checker.best_metric) print(checker(val_d2)) print(checker.best_metric)
294coder/Efficient-MIF
2,092
utils/model_perf_utils.py
import torch import time import pynvml from tqdm import trange def get_gpu_memory_usage(device_idx: int): pynvml.nvmlInit() handle = pynvml.nvmlDeviceGetHandleByIndex(device_idx) info = pynvml.nvmlDeviceGetMemoryInfo(handle) print(f"Total GPU memory: {info.total / 1024**3:.2f} GB") print(f"Used GPU memory: {info.used / 1024**3:.2f} GB") print(f"Free GPU memory: {info.free / 1024**3:.2f} GB") pynvml.nvmlShutdown() def measure_throughput(model: "torch.nn.Module", input_size: "list[tuple[int, ...]] | tuple[int, ...] | None", batch_size: int, num_warmup: int=10, num_iterations: int=50): device = next(model.parameters()).device device_idx = int(str(device)[-1]) get_gpu_memory_usage(device_idx) model.eval() if isinstance(input_size[0], tuple): dummy_input = [] for inp_shape in input_size: if inp_shape is None: inp = torch.randn(batch_size, device=device) else: inp = torch.randn(batch_size, *inp_shape, device=device) dummy_input.append(inp) elif isinstance(input_size, (list, tuple)): dummy_input = [torch.randn(batch_size, *input_size, device=device)] elif input_size is None: dummy_input = [torch.randn(batch_size, device=device)] print(f"Warming up {num_warmup} iterations ...") with torch.no_grad(): for _ in trange(num_warmup): _ = model(*dummy_input) torch.cuda.synchronize() print(f"Measuring model for {num_iterations} iterations ...") total_time = 0 with torch.no_grad(): for _ in trange(num_iterations): start_time = time.time() _ = model(*dummy_input) torch.cuda.synchronize() end_time = time.time() total_time += (end_time - start_time) images_per_second = (num_iterations * batch_size) / total_time print(f"Throughput: {images_per_second:.2f} images/second") return images_per_second
294coder/Efficient-MIF
3,012
utils/print_helper.py
# # For licensing see accompanying LICENSE file. # Copyright (C) 2022 Apple Inc. All Rights Reserved. # # only for printing in color import time from typing import Optional import sys import os text_colors = { "logs": "\033[34m", # 033 is the escape code and 34 is the color code "info": "\033[32m", "warning": "\033[33m", "debug": "\033[93m", "error": "\033[31m", "bold": "\033[1m", "end_color": "\033[0m", "light_red": "\033[36m", } def get_curr_time_stamp() -> str: return time.strftime("%Y-%m-%d %H:%M:%S") def error(message: str) -> None: time_stamp = get_curr_time_stamp() error_str = ( text_colors["error"] + text_colors["bold"] + "ERROR " + text_colors["end_color"] ) print("{} - {} - {}".format(time_stamp, error_str, message), flush=True) print("{} - {} - {}".format(time_stamp, error_str, "Exiting!!!"), flush=True) exit(-1) def color_text(in_text: str) -> str: return text_colors["light_red"] + in_text + text_colors["end_color"] def log(message: str) -> None: time_stamp = get_curr_time_stamp() log_str = ( text_colors["logs"] + text_colors["bold"] + "LOGS " + text_colors["end_color"] ) print("{} - {} - {}".format(time_stamp, log_str, message)) def warning(message: str) -> None: time_stamp = get_curr_time_stamp() warn_str = ( text_colors["warning"] + text_colors["bold"] + "WARNING" + text_colors["end_color"] ) print("{} - {} - {}".format(time_stamp, warn_str, message)) def info(message: str, print_line: Optional[bool] = False) -> None: time_stamp = get_curr_time_stamp() info_str = ( text_colors["info"] + text_colors["bold"] + "INFO " + text_colors["end_color"] ) print("{} - {} - {}".format(time_stamp, info_str, message)) if print_line: double_dash_line(dashes=150) def debug(message: str) -> None: time_stamp = get_curr_time_stamp() log_str = ( text_colors["debug"] + text_colors["bold"] + "DEBUG " + text_colors["end_color"] ) print("{} - {} - {}".format(time_stamp, log_str, message)) def double_dash_line(dashes: Optional[int] = 75) -> None: print(text_colors["error"] + "=" * dashes + text_colors["end_color"]) def singe_dash_line(dashes: Optional[int] = 67) -> None: print("-" * dashes) def print_header(header: str) -> None: double_dash_line() print( text_colors["info"] + text_colors["bold"] + "=" * 50 + str(header) + text_colors["end_color"] ) double_dash_line() def print_header_minor(header: str) -> None: print( text_colors["warning"] + text_colors["bold"] + "=" * 25 + str(header) + text_colors["end_color"] ) def disable_printing(): sys.stdout = open(os.devnull, "w") def enable_printing(): sys.stdout = sys.__stdout__ if __name__ == '__main__': log('12345')
294coder/Efficient-MIF
2,118
utils/save_to_mat.py
import h5py import numpy as np import os import os.path as osp import glob import scipy.io as io from tqdm import tqdm import matplotlib.pyplot as plt def process_mat_to_single_mat(file, show=True, force_all=True): name = file.split('/')[-1].strip('.mat') path = f'/Data2/ZiHanCao/exps/panformer/visualized_img/{name}' print(path) save_prefix = 'output_mulExm_' if not osp.exists(path) or force_all: os.makedirs(path, exist_ok=True) print(f'make dir {name}') else: print('already process this dir') return mat_file = io.loadmat(file) print(f'has keys: {mat_file.keys()}') if 'wv3' in path: rgb_index = [4,2,0] const = 2047 elif 'gf5' in path: rgb_index = [49, 39, 19] const = 1 elif 'gf' in path or 'qb' in path: rgb_index = [2,1,0] const = 1023 elif 'cave' in path or 'harvard'in path: rgb_index = [29,19,9] const = 1 elif 'houston' in path: rgb_index = [39, 29, 19] const = 1 else: rgb_index = [29, 19, 9] const = 1 sr = mat_file.get('sr') if show: ncols = int(np.ceil(sr.shape[0]/4)) fig, axes = plt.subplots(4, ncols, figsize=(ncols*4, 4*4)) axes = axes.flatten() if sr is None: print('has no key sr') else: bar = tqdm(range(sr.shape[0])) for i in bar: save_path = osp.join(path, save_prefix+f'{i}.mat') sr_i = np.transpose(sr[i, ...], [1,2,0]) save_d = {'sr': sr_i} io.savemat(save_path, save_d) bar.set_description(f'save {i}.mat') if show: axes[i].imshow(sr_i[..., rgb_index] / const) axes[i].set_axis_off() if show: plt.tight_layout() fig.savefig(path + '/sr.png') print('---------------'*10) files = glob.glob('/Data2/ZiHanCao/exps/panformer/visualized_img/*.mat') print('found files: ', files) for file in files: process_mat_to_single_mat(file, force_all=True)
2977094657/DidaAPI
972
frontend/docs/api/user/get-user-info.md
# 获取用户信息 获取当前登录用户的基本信息。 ## 接口信息 - **原始API**: `https://api.dida365.com/api/v2/user/profile` - **方法**: `GET` - **认证**: 需要登录认证 ## 请求参数 无需参数,使用当前认证会话。 ## 响应格式 ### 成功响应 ```json { "username": "string", "email": "string", "picture": "string", "locale": "zh_CN", "userCode": "string", "timezone": "Asia/Shanghai", "nickname": "string", "phone": "string", "createdTime": "string", "lastLoginTime": "string", "premium": false, "subscriptionType": "string", "subscriptionExpiry": "string" } ``` ## 响应字段说明 | 字段 | 类型 | 描述 | |------|------|------| | username | string | 用户名 | | email | string | 邮箱地址 | | picture | string | 头像URL | | locale | string | 语言设置 | | userCode | string | 用户代码 | | timezone | string | 时区设置 | | nickname | string | 昵称 | | phone | string | 手机号 | | createdTime | string | 注册时间 | | lastLoginTime | string | 最后登录时间 | | premium | boolean | 是否为高级用户 | | subscriptionType | string | 订阅类型 | | subscriptionExpiry | string | 订阅到期时间 |
2977094657/DidaAPI
2,818
frontend/docs/api/custom/export-tasks-excel.md
# 导出任务到Excel 将用户的所有任务导出为Excel文件,包含全部任务、已完成任务、放弃任务、垃圾桶任务四个工作表。 ## 接口信息 - **接口URL**: `http://localhost:8000/custom/export/tasks/excel` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 本项目自定义接口 ## 功能说明 此接口会将用户的所有任务数据导出为Excel文件,包含以下四个工作表: ### 工作表1:全部任务 包含当前所有任务(未完成和已完成),数据来源于 `/api/v2/batch/check/0` 接口。 ### 工作表2:已完成任务 包含历史已完成的任务,数据来源于 `/api/v2/project/all/closed?status=Completed` 接口。 **分页获取机制**: - 第一次请求不携带`to`参数 - 后续请求使用上次响应最后一个任务的`completedTime`作为`to`参数 - 重复请求直到返回数据少于50条为止,确保获取所有历史数据 ### 工作表3:放弃任务 包含历史放弃的任务,数据来源于 `/api/v2/project/all/closed?status=Abandoned` 接口。 **分页获取机制**: - 第一次请求不携带`to`参数 - 后续请求使用上次响应最后一个任务的`completedTime`作为`to`参数 - 重复请求直到返回数据少于50条为止,确保获取所有历史数据 ### 工作表4:垃圾桶任务 包含已删除的任务,数据来源于 `/api/v2/project/all/trash/page` 接口。 ## 导出字段 每个工作表包含任务的完整字段信息(展平后),包括但不限于: ### 基本信息 - 任务ID - 任务标题 - 任务内容 - 任务描述 - 项目ID - 项目名称 - 排序顺序 ### 状态信息 - 任务状态(文本描述) - 状态代码(数字) - 优先级 - 完成进度 - 删除状态 ### 时间信息 - 创建时间 - 修改时间 - 开始日期 - 截止日期 - 置顶时间 - 完成时间 - 删除时间 ### 时区和时间设置 - 时区 - 是否浮动时间 - 是否全天任务 ### 重复设置 - 重复任务ID - 重复标志 - 重复来源 - 首次重复日期 ### 提醒设置 - 提醒设置 - 提醒列表 - 排除日期 ### 层级关系 - 父任务ID - 子任务ID列表 ### 其他属性 - 标签列表 - 子项目 - 附件数量 - 评论数量 - 列ID - 类型 - 图片模式 - 创建者ID - 删除者ID - 实体标签 - 番茄钟摘要 - 专注摘要 - 附件详情 ## 请求示例 ```http GET http://localhost:8000/custom/export/tasks/excel HTTP/1.1 Host: localhost:8000 Accept: application/vnd.openxmlformats-officedocument.spreadsheetml.sheet ``` ## 响应格式 ### 成功响应 **状态码**: `200 OK` **Content-Type**: `application/vnd.openxmlformats-officedocument.spreadsheetml.sheet` **响应头**: ```http Content-Disposition: attachment; filename=滴答清单任务导出_20250106_143022.xlsx ``` **响应体**: Excel文件的二进制数据 ### 错误响应 **状态码**: `401 Unauthorized` ```json { "detail": "未设置认证会话,请先完成登录" } ``` **状态码**: `500 Internal Server Error` ```json { "detail": "导出失败: 具体错误信息" } ``` ## 使用说明 1. **认证要求**: 必须先完成认证,设置有效的会话 2. **文件格式**: 导出的Excel文件包含四个工作表 3. **文件命名**: 文件名格式为 `滴答清单任务导出_YYYYMMDD_HHMMSS.xlsx` 4. **数据完整性**: 包含任务的所有字段,无遗漏 5. **分页处理**: 自动处理已完成任务和放弃任务的分页,获取所有历史数据 ## 辅助接口 ### 获取导出信息 **接口URL**: `http://localhost:8000/custom/export/tasks/excel/info` **请求方法**: `GET` **功能**: 获取当前用户任务的统计信息,用于导出前预览 **响应示例**: ```json { "auth_status": true, "all_tasks_count": 150, "completed_tasks_count": 89, "abandoned_tasks_count": 15, "trash_tasks_count": 12, "session_info": { "has_session": true, "session_id": "session_123", "created_at": "2025-01-06T14:30:22" } } ``` ## 注意事项 1. 导出过程可能需要一些时间,特别是当任务数量较多时 2. 已完成任务和放弃任务会通过分页获取所有历史数据,确保数据完整性 3. 如果某个数据源获取失败,对应的工作表将为空,但不影响其他工作表 4. 建议在网络状况良好时进行导出操作 5. 导出的Excel文件可以用Microsoft Excel、WPS Office等软件打开 ## 相关接口 - [获取所有任务](../tasks/get-all-tasks.md) - 全部任务数据源 - [获取已完成任务](../tasks/get-completed-tasks.md) - 已完成任务数据源 - [获取已完成/已放弃任务](../tasks/get-completed-tasks.md) - 放弃任务数据源(使用status=Abandoned参数) - [获取垃圾桶任务](../tasks/get-trash-tasks.md) - 垃圾桶任务数据源
2977094657/DidaAPI
3,103
frontend/docs/api/custom/export-focus-excel.md
# 导出专注记录到Excel 将用户的所有专注记录导出为Excel文件,包含完整的专注时间线数据。 ## 接口信息 - **接口URL**: `http://localhost:8000/custom/export/focus/excel` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 本项目自定义接口 ## 功能说明 此接口会将用户的所有专注记录数据导出为Excel文件,包含以下工作表: ### 工作表:专注记录时间线 包含所有专注记录的详细信息,数据来源于 `/api/v2/pomodoros/timeline` 接口,支持自动分页获取所有历史数据。 ## 导出字段 采用紧凑型展示方式,每行一个专注会话,通过特殊格式直观展示专注和暂停模式: ### 主要字段 - **会话ID**: 专注会话的唯一标识符 - **会话时间**: 会话的日期和时间范围,格式:`2025-06-05 11:44 - 17:33` - **总时长**: 会话总时长,格式:`5小时49分钟` - **暂停时长**: 会话中的暂停时长,格式:`1小时11分钟` - **任务标题**: 专注的任务标题(多个任务用分号分隔) - **项目**: 任务所属项目(多个项目用分号分隔) ### 核心展示字段 - **专注时间段**: 详细的专注时间轴,格式示例: ``` 11:44-12:09(25分钟) → [暂停1小时11分钟] → 13:20-17:33(4小时13分钟) ``` - **暂停模式**: 暂停情况的统计分析,格式示例: ``` 暂停1次(1小时11分钟) 暂停3次(总计2小时30分钟, 平均50分钟) 无暂停 ``` ### 分析字段 - **效率(%)**: 实际专注时长占总时长的百分比 - **时间段数量**: 包含的专注时间段数量 - **会话类型**: 会话类型标识 - **实体标签**: 版本控制标签 ### 展示示例 | 会话ID | 会话时间 | 总时长 | 暂停时长 | 任务标题 | 专注时间段 | 暂停模式 | 效率(%) | |--------|----------|--------|----------|----------|------------|----------|---------| | 6841831a... | 2025-06-05 11:44 - 17:33 | 5小时49分钟 | 1小时11分钟 | 从web端抓包所有api | 11:44-12:09(25分钟) → [暂停1小时11分钟] → 13:20-17:33(4小时13分钟) | 暂停1次(1小时11分钟) | 79.7% | | 6841599c... | 2025-06-05 08:47 - 10:11 | 1小时24分钟 | 0分钟 | 从web端抓包所有api | 08:47-10:11(1小时24分钟) | 无暂停 | 100.0% | **重要说明**: - 每个专注会话只占一行,信息紧凑但完整 - "专注时间段"字段能直观显示专注过程中的暂停和继续情况 - 时间格式自动优化显示(秒/分钟/小时) - 暂停模式提供统计分析,便于了解专注习惯 ## 请求示例 ```http GET http://localhost:8000/custom/export/focus/excel HTTP/1.1 Host: localhost:8000 Accept: application/vnd.openxmlformats-officedocument.spreadsheetml.sheet ``` ## 响应格式 ### 成功响应 **状态码**: `200 OK` **Content-Type**: `application/vnd.openxmlformats-officedocument.spreadsheetml.sheet` **响应头**: ```http Content-Disposition: attachment; filename=滴答清单专注记录导出_20250106_143022.xlsx ``` **响应体**: Excel文件的二进制数据 ### 错误响应 **状态码**: `401 Unauthorized` ```json { "detail": "未设置认证会话,请先完成登录" } ``` **状态码**: `500 Internal Server Error` ```json { "detail": "导出失败: 具体错误信息" } ``` ## 分页处理说明 专注记录时间线接口支持分页,本导出功能会自动处理分页: 1. **分页参数**: 使用 `to` 参数进行分页,值为上一页最后一条记录的 `startTime` 2. **时间转换**: 自动将时间字符串转换为中国时间戳(+8小时) 3. **自动分页**: 持续获取直到没有更多数据 4. **数据完整性**: 确保获取所有历史专注记录 ## 使用说明 1. **认证要求**: 必须先完成认证,设置有效的会话 2. **文件格式**: 导出的Excel文件包含专注记录时间线工作表 3. **文件命名**: 文件名格式为 `滴答清单专注记录导出_YYYYMMDD_HHMMSS.xlsx` 4. **数据完整性**: 包含专注记录的所有字段,无遗漏 5. **分页处理**: 自动处理分页,获取所有历史数据 ## 辅助接口 ### 获取导出信息 **接口URL**: `http://localhost:8000/custom/export/focus/excel/info` **请求方法**: `GET` **功能**: 获取当前用户专注记录的统计信息,用于导出前预览 **响应示例**: ```json { "auth_status": true, "focus_records_count_estimate": "31+ (需要分页获取完整数据)", "session_info": { "has_session": true, "session_id": "session_123", "created_at": "2025-01-06T14:30:22" }, "note": "专注记录数量为预估值,实际导出时会获取所有历史数据" } ``` ## 注意事项 1. 导出过程可能需要一些时间,特别是当专注记录数量较多时 2. 会通过分页自动获取所有历史专注记录,确保数据完整性 3. 如果数据源获取失败,导出将失败并返回错误信息 4. 建议在网络状况良好时进行导出操作 5. 导出的Excel文件可以用Microsoft Excel、WPS Office等软件打开 ## 相关接口 - [获取专注记录时间线](../pomodoros/focus-timeline.md) - 专注记录数据源 - [获取专注详情分布](../pomodoros/focus-distribution.md) - 专注分布统计 - [导出任务到Excel](./export-tasks-excel.md) - 任务导出功能
281677160/openwrt-package
3,222
luci-app-passwall/root/usr/share/passwall/tasks.sh
#!/bin/sh ## 循环更新脚本 CONFIG=passwall APP_PATH=/usr/share/$CONFIG TMP_PATH=/tmp/etc/$CONFIG LOCK_FILE=/tmp/lock/${CONFIG}_tasks.lock CFG_UPDATE_INT=0 config_n_get() { local ret=$(uci -q get "${CONFIG}.${1}.${2}" 2>/dev/null) echo "${ret:=$3}" } config_t_get() { local index=${4:-0} local ret=$(uci -q get "${CONFIG}.@${1}[${index}].${2}" 2>/dev/null) echo "${ret:=${3}}" } exec 99>"$LOCK_FILE" flock -n 99 if [ "$?" != 0 ]; then exit 0 fi while true do if [ "$CFG_UPDATE_INT" -ne 0 ]; then stop_week_mode=$(config_t_get global_delay stop_week_mode) stop_interval_mode=$(config_t_get global_delay stop_interval_mode) stop_interval_mode=$(expr "$stop_interval_mode" \* 60) if [ -n "$stop_week_mode" ]; then [ "$stop_week_mode" = "8" ] && { [ "$(expr "$CFG_UPDATE_INT" % "$stop_interval_mode")" -eq 0 ] && /etc/init.d/$CONFIG stop > /dev/null 2>&1 & } fi start_week_mode=$(config_t_get global_delay start_week_mode) start_interval_mode=$(config_t_get global_delay start_interval_mode) start_interval_mode=$(expr "$start_interval_mode" \* 60) if [ -n "$start_week_mode" ]; then [ "$start_week_mode" = "8" ] && { [ "$(expr "$CFG_UPDATE_INT" % "$start_interval_mode")" -eq 0 ] && /etc/init.d/$CONFIG start > /dev/null 2>&1 & } fi restart_week_mode=$(config_t_get global_delay restart_week_mode) restart_interval_mode=$(config_t_get global_delay restart_interval_mode) restart_interval_mode=$(expr "$restart_interval_mode" \* 60) if [ -n "$restart_week_mode" ]; then [ "$restart_week_mode" = "8" ] && { [ "$(expr "$CFG_UPDATE_INT" % "$restart_interval_mode")" -eq 0 ] && /etc/init.d/$CONFIG restart > /dev/null 2>&1 & } fi autoupdate=$(config_t_get global_rules auto_update) weekupdate=$(config_t_get global_rules week_update) hourupdate=$(config_t_get global_rules interval_update) hourupdate=$(expr "$hourupdate" \* 60) if [ "$autoupdate" = "1" ]; then [ "$weekupdate" = "8" ] && { [ "$(expr "$CFG_UPDATE_INT" % "$hourupdate")" -eq 0 ] && lua $APP_PATH/rule_update.lua log all cron > /dev/null 2>&1 & } fi TMP_SUB_PATH=$TMP_PATH/sub_tasks mkdir -p $TMP_SUB_PATH for item in $(uci show ${CONFIG} | grep "=subscribe_list" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do if [ "$(config_n_get $item auto_update 0)" = "1" ]; then cfgid=$(uci show ${CONFIG}.$item | head -n 1 | cut -d '.' -sf 2 | cut -d '=' -sf 1) remark=$(config_n_get $item remark) week_update=$(config_n_get $item week_update) hour_update=$(config_n_get $item interval_update) echo "$cfgid" >> $TMP_SUB_PATH/${week_update}_${hour_update} fi done [ -d "${TMP_SUB_PATH}" ] && { for name in $(ls ${TMP_SUB_PATH}); do week_update=$(echo $name | awk -F '_' '{print $1}') hour_update=$(echo $name | awk -F '_' '{print $2}') hour_update=$(expr "$hour_update" \* 60) cfgids=$(echo -n $(cat ${TMP_SUB_PATH}/${name}) | sed 's# #,#g') [ "$week_update" = "8" ] && { [ "$(expr "$CFG_UPDATE_INT" % "$hour_update")" -eq 0 ] && lua $APP_PATH/subscribe.lua start $cfgids cron > /dev/null 2>&1 & } done rm -rf $TMP_SUB_PATH } fi CFG_UPDATE_INT=$(expr "$CFG_UPDATE_INT" + 10) sleep 600 done 2>/dev/null
294coder/Efficient-MIF
10,627
utils/metric_sharpening.py
import numpy as np import torch from skimage.metrics import structural_similarity, peak_signal_noise_ratio from functools import partial from warnings import warn import sys sys.path.append('./') from utils.misc import to_numpy, dict_to_str from utils._metric_legacy import analysis_accu, indexes_evaluation_FS class NonAnalysis(object): def __init__(self): self.acc_ave = {} # only used as attribution pass def __call__(self, *args, **kwargs): pass def __repr__(self): return 'NonAnalysis()' # TODO: need to be tested the new metric analysis # FIXME: this python code is not same as matlab code, you should use matlab code to get the real accuracy # only used in training and validate class AnalysisPanAcc(object): def __init__(self, ratio=4, ref=True, ergas_ratio: int = 4, **unref_factory_kwargs): """pansharpening metric analysis class Args: ratio (int, optional): fusion ratio. Defaults to 4. ref (bool, optional): reduce-resolution or full-resolution. Defaults to True. ergas_ratio (int, optional): previous api (may decrepated soon). Defaults to 4. unref_factory_kwargs(dict): sensor, default_max_value. Defaults to {'sensor': 'default', 'default_max_value': None}. Raises: ValueError: _description_ """ # ergas_ratio is decrepated if ratio is None: ratio = ergas_ratio warn('@ergas_ratio is deprecated, use ratio instead') self.ratio = ratio self.ref = ref # metric functions if ref: self.__sam_ergas_psnr_cc_one_image = partial(analysis_accu, ratio=ergas_ratio, choices=5) self.ssim = ssim_batch_tensor_metric else: # @sensor in ['QB', 'IKONOS', 'WV2', 'WV3', 'default'] assert 'sensor' in unref_factory_kwargs or 'default_max_value' in unref_factory_kwargs, \ '@sensor or @default_max_value should be specified in unrefactory_kwargs' sensor = unref_factory_kwargs.pop('sensor', 'default').upper() if sensor == 'DEFAULT': warn('sensor is not specified, use default sensor type') self.default_max_value = unref_factory_kwargs.pop('default_max_value', None) if self.default_max_value is None: _default_max_value = {'QB': 2047, 'IKONOS': 1023, 'WV2': 2047, 'WV3': 2047, 'GF2': 1023, 'DEFAULT': 2047, 'CAVE_X4': 1, 'CAVE_X8': 1, 'HARVARD_X': 1, 'HARVARD_X8': 1, 'GF5':1, 'GF2-GF5': 1,} self.default_max_value = _default_max_value.get(sensor) print(f'>>> @default_max_value is not specified, set it according to @sensor:' f'{sensor, self.default_max_value}\n', '-'*20) self.FS_metric_fn = partial(indexes_evaluation_FS, L=11, Qblocks_size=32, sensor=sensor, th_values=0, ratio=ratio, flagQNR=False) # tracking accuracy self._acc_d = {} self._call_n = 0 self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if ref else \ {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} @property def empty_acc(self): return {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \ {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} @staticmethod def permute_dim(*args, permute_dims=(1, 2, 0)): l = [] for i in args: l.append(i.permute(*permute_dims)) return l @staticmethod def _sum_acc(d_ave, d_now, n, n2=1): assert len(d_ave) == len(d_now) for k in d_ave.keys(): v2 = d_now[k] * n2 d_ave[k] *= n d_ave[k] += v2.cpu().item() if isinstance(v2, torch.Tensor) else v2 return d_ave @staticmethod def _average_acc(d_ave, n): for k in d_ave.keys(): d_ave[k] /= n return d_ave def sam_ergas_psnr_cc_batch(self, b_gt, b_pred): n = b_gt.shape[0] # input shape should be [B, C, H, W] acc_ds = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0.} for i, (img1, img2) in enumerate(zip(b_gt, b_pred)): img1, img2 = self.permute_dim(img1, img2) acc_d = self.__sam_ergas_psnr_cc_one_image(img1, img2) acc_ds = self._sum_acc(acc_ds, acc_d, i) acc_ds = self._average_acc(acc_ds, i + 1) return acc_ds def D_lambda_D_s_HQNR_batch(self, sr=None, ms=None, lms=None, pan=None): assert sr is not None and lms is not None and pan is not None and ms is not None if ms is None: ms = torch.nn.functional.interpolate(lms, scale_factor=1/self.rato, mode='bilinear', align_corners=False) acc_ds = {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} sr, ms, lms, pan = self.permute_dim(sr, ms, lms, pan, permute_dims=(0, 2, 3, 1)) sr, ms, lms, pan = to_numpy(sr, ms, lms, pan) _max_value = getattr(self, 'default_max_value') sr, ms, lms, pan = map(lambda x: np.clip(x * _max_value, 0, _max_value), [sr, ms, lms, pan]) for i, (sr_i, ms_i, lms_i, pan_i) in enumerate(zip(sr, ms, lms, pan)): QNR_index, D_lambda, D_S, _ = self.FS_metric_fn(I_F=sr_i, I_MS_LR=ms_i, I_MS=lms_i, I_PAN=pan_i) acc_d = dict(HQNR=QNR_index, D_lambda=D_lambda, D_S=D_S) acc_ds = self._sum_acc(acc_ds, acc_d, i) acc_ds = self._average_acc(acc_ds, i + 1) return acc_ds def once_batch_call(self, **kwargs): if self.ref: acc_d1 = self.sam_ergas_psnr_cc_batch(**kwargs) acc_ssim = self.ssim(**kwargs) acc_d1['SSIM'] = acc_ssim else: acc_d1 = self.D_lambda_D_s_HQNR_batch(**kwargs) self._acc_d = acc_d1 return acc_d1 def _call_check_args_to_kwargs(self, *args): def may_np_to_tensor(d): for k, v in d.items(): if not isinstance(v, torch.Tensor): d[k] = torch.tensor(v, dtype=torch.float32) elif v.dtype != torch.float32: d[k] = v.float() return d if len(args) == 2: assert self.ref, 'ref mode should have 2 args' kwargs = dict(b_gt=args[0], b_pred=args[1]) elif len(args) == 3: assert not self.ref, 'unref mode should have more than 2 args' kwargs = dict(sr=args[0], lms=args[1], pan=args[2]) elif len(args) == 4: assert not self.ref, 'unref mode should have more than 2 args' kwargs = dict(sr=args[0], ms=args[1], lms=args[2], pan=args[3]) else: raise ValueError('args should have 2 or 4 elements') return may_np_to_tensor(kwargs) def __call__(self, *args): """ Args: ref mode: b_gt (torch.Tensor): [b, c, h, w] b_pred (torch.Tensor): [b, c, h, w] unref mode: sr (torch.Tensor): [b, c, h, w] ms (torch.Tensor, optional): [b, c, h/ratio, w/ratio] lms (torch.Tensor): [b, c, h, w] pan (torch.Tensor): [b, c, h, w] """ kwargs = self._call_check_args_to_kwargs(*args) n = args[0].shape[0] self.acc_ave = self._sum_acc(self.acc_ave, self.once_batch_call(**kwargs), self._call_n, n2=n) self.acc_ave = self._average_acc(self.acc_ave, self._call_n + n) self._call_n += n return self.acc_ave def clear_history(self, verbose=False): if verbose: print('>> AccAnalysis: clear history') self._acc_d = {} self._call_n = 0 self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \ {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} def print_str(self, decimals=6): return dict_to_str(self.acc_ave, decimals=decimals) def result_str(self): return self.print_str() def __repr__(self) -> str: repr_str = f'AnalysisPanAcc(ratio={self.ratio}, ref={self.ref}):' repr_str += f'\n{self.print_str()}' return repr_str def normalize_to_01(x): # normalize tensor to [0, 1] if isinstance(x, torch.Tensor): x -= x.flatten(-2).min(-1, keepdim=True)[0][..., None] x /= x.flatten(-2).max(-1, keepdim=True)[0][..., None] elif isinstance(x, np.ndarray): x -= x.min((-2, -1), keepdims=True) x /= x.max((-2, -1), keepdims=True) return x def psnr_one_img(img_gt, img_test): """ calculate PSNR for one image :param img_gt: ground truth image, numpy array, shape [H, W, C] :param img_test: test or inference image, numpy array, shape [H, W, C] :return: PSNR, float type """ assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size' return peak_signal_noise_ratio(img_gt, img_test) def psnr_batch_tensor_metric(b_gt, b_pred): """ calculate PSNR for batch tensor images :param b_gt: tensor, shape [B, C, H, W] :param b_test: tensor, shape [B, C, H, W] :return: """ assert b_gt.shape[0] == b_pred.shape[0] bs = b_gt.shape[0] psnr = 0. for gt, t in zip(b_gt, b_pred): psnr += psnr_one_img(*(to_numpy(gt, t))) return psnr / bs def ssim_one_image(img_gt, img_test, channel_axis=0): assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size' return structural_similarity(img_gt, img_test, channel_axis=channel_axis, data_range=1.) def ssim_batch_tensor_metric(b_gt, b_pred): assert b_gt.shape[0] == b_pred.shape[0] bs = b_gt.shape[0] ssim = 0. for gt, t in zip(b_gt, b_pred): ssim += ssim_one_image(*(to_numpy(gt, t)), channel_axis=0) return ssim / bs if __name__ == '__main__': sr = torch.rand(4, 3, 256, 256) ms = torch.rand(4, 3, 64, 64) lms = torch.rand(4, 3, 256, 256) pan = torch.rand(4, 3, 256, 256) gt = torch.rand(4, 3, 256, 256) analysis = AnalysisPanAcc(ref=False, ratio=4, default_max_value=2047) for i in range(2): analysis(sr[i:i+2], ms[i:i+2], lms[i:i+2], pan[i:i+2]) print(analysis.print_str())
294coder/Efficient-MIF
8,375
utils/load_params.py
from functools import partial from numpy import isin import torch import logging from collections import OrderedDict from safetensors.torch import load_file from packaging import version from utils.log_utils import easy_logger def module_load(path, model, device, ddp_rank=None, strict=True, spec_key='shadow_params', logger=None, full_unmatched_log=True): if logger is None: logger = easy_logger(func_name=__name__) model = model.to(device if ddp_rank is None else ddp_rank) place = device if ddp_rank is None else {'cuda:%d' % 0: 'cuda:%d' % ddp_rank} if isinstance(place, torch.device): place = str(place) if path.endswith('pth') or path.endswith('pt'): if version.parse(torch.__version__) >= version.parse('2.4.0'): load_engine = partial(torch.load, map_location=place, weights_only=False) else: load_engine = partial(torch.load, map_location=place) elif path.endswith('safetensors'): load_engine = lambda weight_path, map_location: OrderedDict(load_file(weight_path, device=map_location)) else: raise ValueError try: params = load_engine(path, map_location=place) except Exception: # TODO: the exception is not used logger.print('>>> did not find the pth file, try to find in used_weights/ and ununsed_weights/...', logging.INFO) try: path_used = path.replace('weight/', 'weight/used_weights/') params = load_engine(path_used, map_location=place) except Exception: path_ununsed = path.replace('weight/', 'weight/unused_weights/') params = load_engine(path_ununsed, map_location=place) # parse key if spec_key is not None: parsed_keys = spec_key.split('.') try: for k in parsed_keys: params = params[k] except KeyError: logger.warning(f'>>> not found parsed model `{spec_key}`, load the model directly \n \n') _load_fail_flag = False params_load = params # may be tedious but useful and safe to avoid 'module.' prefix caused error if not strict: logger.warning('model load strict is False, set it to True if you know what you are doing') def _iter_params_load_fn(model, params_load, strict): nonlocal _load_fail_flag if not isinstance(params_load, (list, tuple)): param_load_ziped = list(params_load.items()) if isinstance(params_load, dict) else params_load for (s_name, s_param), (name, param) in zip(param_load_ziped, model.named_parameters()): saved_shape = tuple(s_param.data.shape) required_shape = tuple(param.data.shape) if saved_shape != required_shape: if strict: logger.print( f'param shape unmatched, {name} requires: {required_shape}, but got {s_name}: {saved_shape}', logging.WARNING ) if not full_unmatched_log: logger.print('model load failed! shape of params does not match!', logging.ERROR) raise RuntimeError('model load failed! shape of params does not match!') else: _load_fail_flag = True continue else: logger.print(f'skip the shape mismatched param, param name {name}, ' + f'current shape {required_shape} but loaded shape {saved_shape}', logging.WARNING) continue param.data.copy_(s_param.data) else: for s_param, param in zip(params_load, model.parameters()): required_shape = tuple(param.data.shape) saved_shape = tuple(s_param.data.shape) if saved_shape != required_shape: if strict: logger.print( f'param shape unmatched, requires: {required_shape}, but got {saved_shape}', logging.WARNING ) if not full_unmatched_log: logger.print('model load failed! shape of params does not match!', logging.ERROR) raise RuntimeError('model load failed! shape of params does not match!') else: _load_fail_flag = True continue else: logger.print(f'skip the shape mismatched param, current shape {required_shape} but loaded shape {saved_shape}', logging.WARNING) continue param.data.copy_(s_param.data) def _load_fn(model, params_load, strict): if isinstance(params_load, OrderedDict): # ordered dict model.load_state_dict(params_load, strict=strict) else: _iter_params_load_fn(model, params_load, strict) _load_fn(model, params_load, strict) if _load_fail_flag: raise RuntimeError('model load failed! shape of params does not match!') # except Exception: # # data parallel mode will save params with keys' prefix is 'module'. # odict = {} # for k, v in params_load.items(): # odict['module.' + k] = v # params_load[spec_key] = odict # if 'ema' not in spec_key: # _load_fn(model, params_load, strict) # else: # raise RuntimeError('ema model load failed! shape of params does not match!') logger.print('load pretrain weights', logging.INFO) return model # def resume_load(path, # model, # optim, # lr_scheduler, # ema_model: ExponentialMovingAverage=None, # specific_resume_lr: float = None, # specific_epochs: int = None, # wd_scheduler=None, # device='cuda:0', # ddp_rank=None, # ddp=False): # # @specific_resume_lr(warning: not recommended): # # manually specify learning rate when the lr from last break is too low to update model # # @specific_epochs(warning: not recommended): # # manually specify total epochs when resuming training # model.to(device if ddp_rank is None else ddp_rank) # # assume saved params always on cuda:0 # params = torch.load(path, map_location=device if ddp_rank is None else {'cuda:%d' % 0: 'cuda:%d' % ddp_rank}) # # NOTE: ddp mode will save params with keys' prefix is 'module'. # # now I remove the prefix for just one card circumstance but it conflict with ddp mode. # if ddp: # odict = OrderedDict() # for k, v in params['model'].items(): # odict['module.' + k] = v # params['model'] = odict # model.load_state_dict(params['model']) # if ema_model is not None: # ema_model.load_state_dict(params['ema_model']) # # NOTE: Pytorch 1.12.0 may cause CUDA error in optimizer reloading. see more at # # https://github.com/pytorch/pytorch/issues/80809#issuecomment-1175211598 # optim.load_state_dict(params['optim']) # if specific_resume_lr is not None: # optim.param_groups[0]['lr'] = specific_resume_lr # lr_scheduler.load_state_dict(params['lr_scheduler']) # if specific_epochs is not None: # # FIXME: only support CosineAnnealing lr_scheduler # lr_scheduler.T_max = specific_epochs # resume_ep = params['epochs'] # print(f"last training resume! best metrics are {params['metrics']}") # # warning: if you change total epochs in the resume run, the lr_scheduler may not update lr # if wd_scheduler is not None: # wd_scheduler.load_state_dict(params['wd_scheduler']) # return model, optim, lr_scheduler, wd_scheduler, resume_ep # else: # return model, optim, lr_scheduler, resume_ep
2977094657/DidaAPI
2,234
frontend/docs/api/statistics/task-statistics.md
# 获取任务统计信息 ## 接口信息 - **接口路径**: `GET /statistics/tasks` - **接口描述**: 获取指定日期范围内的任务统计信息 - **请求方式**: GET - **认证要求**: 需要先完成微信登录获取认证会话 ## 请求参数 ### Query参数 | 参数名 | 类型 | 必填 | 描述 | 示例值 | |--------|------|------|------|--------| | start_date | string | 是 | 开始日期,格式: YYYYMMDD | `20231201` | | end_date | string | 是 | 结束日期,格式: YYYYMMDD | `20231207` | ## 请求示例 ```bash curl -X GET "http://localhost:8000/statistics/tasks?start_date=20231201&end_date=20231207" ``` ## 响应格式 ### 成功响应 ```json { "overdueCompletedCount": 2, "onTimeCompletedCount": 15, "noDateCompletedCount": 8, "uncompletedCount": 5, "projectStats": [ { "projectId": "project123", "projectName": "工作项目", "completedCount": 10 }, { "projectId": "project456", "projectName": "个人项目", "completedCount": 5 } ], "tagStats": [ { "tagName": "重要", "completedCount": 8 }, { "tagName": "紧急", "completedCount": 4 } ] } ``` ### 错误响应 #### 参数错误 ```json { "detail": [ { "loc": ["query", "start_date"], "msg": "field required", "type": "value_error.missing" } ] } ``` #### 未认证 ```json { "error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录" } ``` #### 服务错误 ```json { "error": "service_error", "message": "获取任务统计信息失败,请稍后重试" } ``` ## 响应字段说明 | 字段名 | 类型 | 描述 | |--------|------|------| | overdueCompletedCount | integer | 逾期完成任务数 | | onTimeCompletedCount | integer | 按时完成任务数 | | noDateCompletedCount | integer | 无日期任务完成数 | | uncompletedCount | integer | 未完成任务数 | | projectStats | array | 按项目的完成统计 | | projectStats[].projectId | string | 项目ID | | projectStats[].projectName | string | 项目名称 | | projectStats[].completedCount | integer | 该项目完成任务数 | | tagStats | array | 按标签的完成统计 | | tagStats[].tagName | string | 标签名称 | | tagStats[].completedCount | integer | 该标签完成任务数 | ## 使用说明 1. **认证要求**: 需要先调用微信登录接口获取认证会话 2. **日期格式**: 日期必须使用YYYYMMDD格式 3. **统计维度**: 提供多维度的任务完成统计 4. **时间范围**: 支持自定义时间范围查询 ## 相关接口 - [获取通用统计信息](./general-statistics.md) - [获取用户排名统计](../statistics.md) - [获取任务统计](../tasks/get-tasks-summary.md) - [微信登录流程](../auth/wechat-login-flow.md) ## 注意事项 - 需要先完成微信登录获取认证会话 - 日期格式必须正确,否则会返回参数错误 - 统计数据基于指定时间范围内的任务 - 项目和标签统计按完成数量降序排列
2977094657/DidaAPI
552
frontend/docs/api/statistics/index.md
# 统计分析接口 本模块提供了滴答清单的各种统计分析功能,帮助用户了解自己的任务完成情况和使用习惯。 ## 接口列表 ### 用户排名统计 - [获取用户排名统计](../statistics.md) - 获取用户在滴答清单中的排名和基本统计信息 ### 通用统计 - [获取通用统计信息](./general-statistics.md) - 获取概览、成就值、趋势等通用统计信息 ### 任务统计 - [获取任务统计信息](./task-statistics.md) - 获取指定日期范围内的任务统计信息 ## 使用说明 1. **认证要求**: 所有统计接口都需要先完成微信登录获取认证会话 2. **数据实时性**: 统计数据实时更新,反映最新的用户活动 3. **时间范围**: 部分接口支持自定义时间范围查询 4. **多维度统计**: 提供任务、项目、标签等多维度的统计分析 ## 相关模块 - [任务管理](../tasks/get-all-tasks.md) - 任务相关操作 - [习惯管理](../habits.md) - 习惯打卡统计 - [番茄专注](../pomodoros.md) - 专注时间统计 - [认证相关](../auth/) - 登录认证流程
281677160/openwrt-package
23,981
luci-app-passwall/root/usr/share/passwall/helper_smartdns_add.lua
require "luci.sys" local api = require "luci.passwall.api" local appname = "passwall" local var = api.get_args(arg) local FLAG = var["-FLAG"] local SMARTDNS_CONF = var["-SMARTDNS_CONF"] local LOCAL_GROUP = var["-LOCAL_GROUP"] local REMOTE_GROUP = var["-REMOTE_GROUP"] local REMOTE_PROXY_SERVER = var["-REMOTE_PROXY_SERVER"] local USE_DEFAULT_DNS = var["-USE_DEFAULT_DNS"] local REMOTE_DNS = var["-REMOTE_DNS"] local TUN_DNS = var["-TUN_DNS"] local DNS_MODE = var["-DNS_MODE"] local REMOTE_FAKEDNS = var["-REMOTE_FAKEDNS"] local TCP_NODE = var["-TCP_NODE"] local USE_DIRECT_LIST = var["-USE_DIRECT_LIST"] local USE_PROXY_LIST = var["-USE_PROXY_LIST"] local USE_BLOCK_LIST = var["-USE_BLOCK_LIST"] local USE_GFW_LIST = var["-USE_GFW_LIST"] local CHN_LIST = var["-CHN_LIST"] local DEFAULT_PROXY_MODE = var["-DEFAULT_PROXY_MODE"] local NO_PROXY_IPV6 = var["-NO_PROXY_IPV6"] local NO_LOGIC_LOG = var["-NO_LOGIC_LOG"] local NFTFLAG = var["-NFTFLAG"] local SUBNET = var["-SUBNET"] local uci = api.uci local sys = api.sys local fs = api.fs local datatypes = api.datatypes local TMP_PATH = "/tmp/etc/" .. appname local TMP_ACL_PATH = TMP_PATH .. "/acl" local RULES_PATH = "/usr/share/" .. appname .. "/rules" local FLAG_PATH = TMP_ACL_PATH .. "/" .. FLAG local TMP_CONF_FILE = FLAG_PATH .. "/smartdns.conf" local config_lines = {} local tmp_lines = {} local USE_GEOVIEW = uci:get(appname, "@global_rules[0]", "enable_geoview") local function log(...) if NO_LOGIC_LOG == "1" then return end api.log(...) end local function is_file_nonzero(path) if path and #path > 1 then if sys.exec('[ -s "%s" ] && echo -n 1' % path) == "1" then return true end end return nil end local function insert_unique(dest_table, value, lookup_table) if not lookup_table[value] then table.insert(dest_table, value) lookup_table[value] = true end end local function merge_array(array1, array2) for i, line in ipairs(array2) do table.insert(array1, #array1 + 1, line) end end local function insert_array_before(array1, array2, target) --将array2插入到array1的target前面,target不存在则追加 for i, line in ipairs(array1) do if line == target then for j = #array2, 1, -1 do table.insert(array1, i, array2[j]) end return end end merge_array(array1, array2) end local function insert_array_after(array1, array2, target) --将array2插入到array1的target后面,target不存在则追加 for i, line in ipairs(array1) do if line == target then for j = 1, #array2 do table.insert(array1, i + j, array2[j]) end return end end merge_array(array1, array2) end local function get_geosite(list_arg, out_path) local geosite_path = uci:get(appname, "@global_rules[0]", "v2ray_location_asset") or "/usr/share/v2ray/" geosite_path = geosite_path:match("^(.*)/") .. "/geosite.dat" if not is_file_nonzero(geosite_path) then return 1 end if api.is_finded("geoview") and list_arg and out_path then sys.exec("geoview -type geosite -append=true -input " .. geosite_path .. " -list '" .. list_arg .. "' -output " .. out_path) return 0 end return 1 end if not fs.access(FLAG_PATH) then fs.mkdir(FLAG_PATH) end local LOCAL_EXTEND_ARG = "" if LOCAL_GROUP == "nil" then LOCAL_GROUP = nil log(" * 注意:国内分组名未设置,可能会导致 DNS 分流错误!") else --从smartdns配置中读取参数 local custom_conf_path = "/etc/smartdns/custom.conf" local options = { {key = "dualstack_ip_selection", config_key = "dualstack-ip-selection", yes_no = true, arg_yes = "-d yes", arg_no = "-d no", default = "yes"}, {key = "speed_check_mode", config_key = "speed-check-mode", prefix = "-c ", default = "ping,tcp:80,tcp:443"}, {key = "serve_expired", config_key = "serve-expired", yes_no = true, arg_yes = "", arg_no = "-no-serve-expired", default = "yes"}, {key = "response_mode", config_key = "response-mode", prefix = "-r ", default = "first-ping"}, {key = "rr_ttl", config_key = "rr-ttl", prefix = "-rr-ttl "}, {key = "rr_ttl_min", config_key = "rr-ttl-min", prefix = "-rr-ttl-min "}, {key = "rr_ttl_max", config_key = "rr-ttl-max", prefix = "-rr-ttl-max "} } -- 从 custom.conf 中读取值,以最后出现的值为准 local custom_config = {} local f_in = io.open(custom_conf_path, "r") if f_in then for line in f_in:lines() do line = api.trim(line) if line ~= "" and not line:match("^#") then local param, value = line:match("^(%S+)%s+(%S+)$") if param and value then custom_config[param] = value end end end f_in:close() end -- 从 smartdns 配置中读取值,优先级以 custom.conf 为准 for _, opt in ipairs(options) do local val = custom_config[opt.config_key] or uci:get("smartdns", "@smartdns[0]", opt.key) or opt.default if val == "yes" then val = "1" elseif val == "no" then val = "0" end if opt.yes_no then local arg = (val == "1" and opt.arg_yes or opt.arg_no) if arg and arg ~= "" then LOCAL_EXTEND_ARG = LOCAL_EXTEND_ARG .. (LOCAL_EXTEND_ARG ~= "" and " " or "") .. arg end else if val and (not opt.value or (opt.invert and val ~= opt.value) or (not opt.invert and val == opt.value)) then LOCAL_EXTEND_ARG = LOCAL_EXTEND_ARG .. (LOCAL_EXTEND_ARG ~= "" and " " or "") .. (opt.prefix or "") .. (opt.arg or val) end end end end if not REMOTE_GROUP or REMOTE_GROUP == "nil" then REMOTE_GROUP = "passwall_proxy" if REMOTE_DNS then REMOTE_DNS = REMOTE_DNS:gsub("#", ":") end sys.call('sed -i "/passwall/d" /etc/smartdns/custom.conf >/dev/null 2>&1') end local force_https_soa = uci:get(appname, "@global[0]", "force_https_soa") or 1 local proxy_server_name = "passwall-proxy-server" config_lines = { tonumber(force_https_soa) == 1 and "force-qtype-SOA 65" or "force-qtype-SOA -,65", "server 114.114.114.114 -bootstrap-dns", DNS_MODE == "socks" and string.format("proxy-server socks5://%s -name %s", REMOTE_PROXY_SERVER, proxy_server_name) or nil } if DNS_MODE == "socks" then for w in string.gmatch(REMOTE_DNS, '[^|]+') do local server_dns = api.trim(w) local server_param local dnsType = string.match(server_dns, "^(.-)://") dnsType = dnsType and string.lower(dnsType) or nil local dnsServer = string.match(server_dns, "://(.+)") or server_dns if dnsType and dnsType ~= "" and dnsType ~= "udp" then if dnsType == "tcp" then server_param = "server-tcp " .. dnsServer elseif dnsType == "tls" then server_param = "server-tls " .. dnsServer elseif dnsType == "quic" then server_param = "server-quic " .. dnsServer elseif dnsType == "https" or dnsType == "h3" then local http_host = nil local url = w local port = 443 local s = api.split(w, ",") if s and #s > 1 then url = s[1] local dns_ip = s[2] local host_port = api.get_domain_from_url(s[1]) if host_port and #host_port > 0 then http_host = host_port local s2 = api.split(host_port, ":") if s2 and #s2 > 1 then http_host = s2[1] port = s2[2] end url = url:gsub(http_host, dns_ip) end end server_dns = url if http_host then server_dns = server_dns .. " -http-host " .. http_host end server_param = (dnsType == "https" and "server-https " or "server-h3 ") .. server_dns end else server_param = "server " .. dnsServer end if not api.is_local_ip(w) then server_param = server_param .. " -proxy " .. proxy_server_name end server_param = server_param .. " -group " .. REMOTE_GROUP .. " -exclude-default-group" if SUBNET and SUBNET ~= "" and SUBNET ~= "0" then server_param = server_param .. " -subnet " .. SUBNET end table.insert(config_lines, server_param) end REMOTE_FAKEDNS = 0 else local server_param = string.format("server %s -group %s -exclude-default-group", TUN_DNS:gsub("#", ":"), REMOTE_GROUP) table.insert(config_lines, server_param) log(" - " .. DNS_MODE:gsub("^%l",string.upper) .. " " .. TUN_DNS .. " -> " .. REMOTE_GROUP) end --设置默认 DNS 分组(托底组) local DEFAULT_DNS_GROUP = (USE_DEFAULT_DNS == "direct" and LOCAL_GROUP) or (USE_DEFAULT_DNS == "remote" and REMOTE_GROUP) local only_global = (DEFAULT_PROXY_MODE == "proxy" and CHN_LIST == "0" and USE_GFW_LIST == "0") and 1 --没有启用中国列表和GFW列表时(全局) if only_global == 1 then DEFAULT_DNS_GROUP = REMOTE_GROUP end if DEFAULT_DNS_GROUP then local domain_rules_str = "domain-rules /./ -nameserver " .. DEFAULT_DNS_GROUP if DEFAULT_DNS_GROUP == REMOTE_GROUP then domain_rules_str = domain_rules_str .. " -speed-check-mode none -d no -no-serve-expired" if NO_PROXY_IPV6 == "1" then domain_rules_str = domain_rules_str .. " -address #6" end elseif DEFAULT_DNS_GROUP == LOCAL_GROUP then domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "") end table.insert(config_lines, domain_rules_str) end local setflag = (NFTFLAG == "1") and "inet#passwall#" or "" local set_type = (NFTFLAG == "1") and "-nftset" or "-ipset" --预设排序标签(越往后优先级越高) for i = 1, 8 do table.insert(config_lines, "#--" .. i) end --屏蔽列表 local file_block_host = TMP_ACL_PATH .. "/block_host" if USE_BLOCK_LIST == "1" and not fs.access(file_block_host) then local block_domain, lookup_block_domain = {}, {} local geosite_arg = "" for line in io.lines(RULES_PATH .. "/block_host") do if not line:find("#") and line:find("geosite:") then line = string.match(line, ":([^:]+)$") geosite_arg = geosite_arg .. (geosite_arg ~= "" and "," or "") .. line else line = api.get_std_domain(line) if line ~= "" and not line:find("#") then insert_unique(block_domain, line, lookup_block_domain) end end end if #block_domain > 0 then local f_out = io.open(file_block_host, "w") for i = 1, #block_domain do f_out:write(block_domain[i] .. "\n") end f_out:close() end if USE_GEOVIEW == "1" and geosite_arg ~= "" and api.is_finded("geoview") then if get_geosite(geosite_arg, file_block_host) == 0 then log(" * 解析[屏蔽列表] Geosite 到屏蔽域名表(blocklist)完成") else log(" * 解析[屏蔽列表] Geosite 到屏蔽域名表(blocklist)失败!") end end end if USE_BLOCK_LIST == "1" and is_file_nonzero(file_block_host) then local domain_set_name = "passwall-block" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, file_block_host), string.format("domain-rules /domain-set:%s/ -a #", domain_set_name) } insert_array_after(config_lines, tmp_lines, "#--7") end --始终用国内DNS解析节点域名 local file_vpslist = TMP_ACL_PATH .. "/vpslist" if not is_file_nonzero(file_vpslist) then local f_out = io.open(file_vpslist, "w") local written_domains = {} uci:foreach(appname, "nodes", function(t) local function process_address(address) if address == "engage.cloudflareclient.com" then return end if datatypes.hostname(address) and not written_domains[address] then f_out:write(address .. "\n") written_domains[address] = true end end process_address(t.address) process_address(t.download_address) end) f_out:close() end if is_file_nonzero(file_vpslist) then local domain_set_name = "passwall-vpslist" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, file_vpslist) } local sets = { "#4:" .. setflag .. "passwall_vps", "#6:" .. setflag .. "passwall_vps6" } local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "") domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",") domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "") table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--8") log(string.format(" - 节点列表中的域名(vpslist)使用分组:%s", LOCAL_GROUP or "默认")) end --直连(白名单)列表 local file_direct_host = TMP_ACL_PATH .. "/direct_host" if USE_DIRECT_LIST == "1" and not fs.access(file_direct_host) then local direct_domain, lookup_direct_domain = {}, {} local geosite_arg = "" for line in io.lines(RULES_PATH .. "/direct_host") do if not line:find("#") and line:find("geosite:") then line = string.match(line, ":([^:]+)$") geosite_arg = geosite_arg .. (geosite_arg ~= "" and "," or "") .. line else line = api.get_std_domain(line) if line ~= "" and not line:find("#") then insert_unique(direct_domain, line, lookup_direct_domain) end end end if #direct_domain > 0 then local f_out = io.open(file_direct_host, "w") for i = 1, #direct_domain do f_out:write(direct_domain[i] .. "\n") end f_out:close() end if USE_GEOVIEW == "1" and geosite_arg ~= "" and api.is_finded("geoview") then if get_geosite(geosite_arg, file_direct_host) == 0 then log(" * 解析[直连列表] Geosite 到域名白名单(whitelist)完成") else log(" * 解析[直连列表] Geosite 到域名白名单(whitelist)失败!") end end end if USE_DIRECT_LIST == "1" and is_file_nonzero(file_direct_host) then local domain_set_name = "passwall-directlist" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, file_direct_host) } local sets = { "#4:" .. setflag .. "passwall_white", "#6:" .. setflag .. "passwall_white6" } local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "") domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",") domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "") table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--6") log(string.format(" - 域名白名单(whitelist)使用分组:%s", LOCAL_GROUP or "默认")) end --代理(黑名单)列表 local file_proxy_host = TMP_ACL_PATH .. "/proxy_host" if USE_PROXY_LIST == "1" and not fs.access(file_proxy_host) then local proxy_domain, lookup_proxy_domain = {}, {} local geosite_arg = "" for line in io.lines(RULES_PATH .. "/proxy_host") do if not line:find("#") and line:find("geosite:") then line = string.match(line, ":([^:]+)$") geosite_arg = geosite_arg .. (geosite_arg ~= "" and "," or "") .. line else line = api.get_std_domain(line) if line ~= "" and not line:find("#") then insert_unique(proxy_domain, line, lookup_proxy_domain) end end end if #proxy_domain > 0 then local f_out = io.open(file_proxy_host, "w") for i = 1, #proxy_domain do f_out:write(proxy_domain[i] .. "\n") end f_out:close() end if USE_GEOVIEW == "1" and geosite_arg ~= "" and api.is_finded("geoview") then if get_geosite(geosite_arg, file_proxy_host) == 0 then log(" * 解析[代理列表] Geosite 到代理域名表(blacklist)完成") else log(" * 解析[代理列表] Geosite 到代理域名表(blacklist)失败!") end end end if USE_PROXY_LIST == "1" and is_file_nonzero(file_proxy_host) then local domain_set_name = "passwall-proxylist" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, file_proxy_host) } local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP) domain_rules_str = domain_rules_str .. " -speed-check-mode none" domain_rules_str = domain_rules_str .. " -no-serve-expired" local sets = { "#4:" .. setflag .. "passwall_black" } if NO_PROXY_IPV6 == "1" then domain_rules_str = domain_rules_str .. " -address #6" domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str else table.insert(sets, "#6:" .. setflag .. "passwall_black6") domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str end table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--5") log(string.format(" - 代理域名表(blacklist)使用分组:%s", REMOTE_GROUP or "默认")) end --GFW列表 if USE_GFW_LIST == "1" and is_file_nonzero(RULES_PATH .. "/gfwlist") then local domain_set_name = "passwall-gfwlist" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, RULES_PATH .. "/gfwlist") } local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP) domain_rules_str = domain_rules_str .. " -speed-check-mode none" domain_rules_str = domain_rules_str .. " -no-serve-expired" local sets = { "#4:" .. setflag .. "passwall_gfw" } if NO_PROXY_IPV6 == "1" then domain_rules_str = domain_rules_str .. " -address #6" domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str else table.insert(sets, "#6:" .. setflag .. "passwall_gfw6") domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str end table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--1") log(string.format(" - 防火墙域名表(gfwlist)使用分组:%s", REMOTE_GROUP or "默认")) end --中国列表 if CHN_LIST ~= "0" and is_file_nonzero(RULES_PATH .. "/chnlist") then local domain_set_name = "passwall-chnlist" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, RULES_PATH .. "/chnlist") } if CHN_LIST == "direct" then local sets = { "#4:" .. setflag .. "passwall_chn", "#6:" .. setflag .. "passwall_chn6" } local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "") domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",") domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "") table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--2") log(string.format(" - 中国域名表(chnroute)使用分组:%s", LOCAL_GROUP or "默认")) end --回中国模式 if CHN_LIST == "proxy" then local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP) domain_rules_str = domain_rules_str .. " -speed-check-mode none" domain_rules_str = domain_rules_str .. " -no-serve-expired" local sets = { "#4:" .. setflag .. "passwall_chn" } if NO_PROXY_IPV6 == "1" then domain_rules_str = domain_rules_str .. " -address #6" domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str else table.insert(sets, "#6:" .. setflag .. "passwall_chn6") domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str end table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--2") log(string.format(" - 中国域名表(chnroute)使用分组:%s", REMOTE_GROUP or "默认")) end end --分流规则 if uci:get(appname, TCP_NODE, "protocol") == "_shunt" then local white_domain, lookup_white_domain = {}, {} local shunt_domain, lookup_shunt_domain = {}, {} local file_white_host = FLAG_PATH .. "/shunt_direct_host" local file_shunt_host = FLAG_PATH .. "/shunt_proxy_host" local geosite_white_arg, geosite_shunt_arg = "", "" local t = uci:get_all(appname, TCP_NODE) local default_node_id = t["default_node"] or "_direct" uci:foreach(appname, "shunt_rules", function(s) local _node_id = t[s[".name"]] if _node_id and _node_id ~= "_blackhole" then if _node_id == "_default" then _node_id = default_node_id end local domain_list = s.domain_list or "" for line in string.gmatch(domain_list, "[^\r\n]+") do if line ~= "" and not line:find("#") and not line:find("regexp:") and not line:find("ext:") then if line:find("geosite:") then line = string.match(line, ":([^:]+)$") if _node_id == "_direct" then geosite_white_arg = geosite_white_arg .. (geosite_white_arg ~= "" and "," or "") .. line else geosite_shunt_arg = geosite_shunt_arg .. (geosite_shunt_arg ~= "" and "," or "") .. line end else if line:find("domain:") or line:find("full:") then line = string.match(line, ":([^:]+)$") end line = api.get_std_domain(line) if line ~= "" and not line:find("#") then if _node_id == "_direct" then insert_unique(white_domain, line, lookup_white_domain) else insert_unique(shunt_domain, line, lookup_shunt_domain) end end end end end if _node_id ~= "_direct" then log(string.format(" - Sing-Box/Xray分流规则(%s)使用分组:%s", s.remarks, REMOTE_GROUP or "默认")) end end end) if is_file_nonzero(file_white_host) == nil then if #white_domain > 0 then local f_out = io.open(file_white_host, "w") for i = 1, #white_domain do f_out:write(white_domain[i] .. "\n") end f_out:close() end end if is_file_nonzero(file_shunt_host) == nil then if #shunt_domain > 0 then local f_out = io.open(file_shunt_host, "w") for i = 1, #shunt_domain do f_out:write(shunt_domain[i] .. "\n") end f_out:close() end end if USE_GFW_LIST == "1" and CHN_LIST == "0" and USE_GEOVIEW == "1" and api.is_finded("geoview") then --仅GFW模式解析geosite local return_white, return_shunt if geosite_white_arg ~= "" then return_white = get_geosite(geosite_white_arg, file_white_host) end if geosite_shunt_arg ~= "" then return_shunt = get_geosite(geosite_shunt_arg, file_shunt_host) end if (return_white == nil or return_white == 0) and (return_shunt == nil or return_shunt == 0) then log(" * 解析[分流节点] Geosite 完成") else log(" * 解析[分流节点] Geosite 失败!") end end if is_file_nonzero(file_white_host) then local domain_set_name = "passwall-whitehost" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, file_white_host) } local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "") if USE_DIRECT_LIST == "1" then local sets = { "#4:" .. setflag .. "passwall_white", "#6:" .. setflag .. "passwall_white6" } domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",") else local sets = { "#4:" .. setflag .. "passwall_shunt", "#6:" .. setflag .. "passwall_shunt6" } domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",") end domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "") table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--4") end if is_file_nonzero(file_shunt_host) then local domain_set_name = "passwall-shuntlist" tmp_lines = { string.format("domain-set -name %s -file %s", domain_set_name, file_shunt_host) } local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP) domain_rules_str = domain_rules_str .. " -speed-check-mode none" domain_rules_str = domain_rules_str .. " -no-serve-expired" local sets = { "#4:" .. setflag .. "passwall_shunt" } if NO_PROXY_IPV6 == "1" then domain_rules_str = domain_rules_str .. " -address #6" domain_rules_str = (not only_global and REMOTE_FAKEDNS == "1") and domain_rules_str or (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")) else table.insert(sets, "#6:" .. setflag .. "passwall_shunt6") domain_rules_str = (not only_global and REMOTE_FAKEDNS == "1") and domain_rules_str or (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ",")) end table.insert(tmp_lines, domain_rules_str) insert_array_after(config_lines, tmp_lines, "#--3") end end if #config_lines > 0 then local f_out = io.open(TMP_CONF_FILE, "w") for i = 1, #config_lines do line = config_lines[i] if line ~= "" and not line:find("^#--") then f_out:write(line .. "\n") end end f_out:close() end if DEFAULT_DNS_GROUP then log(string.format(" - 默认 DNS 分组:%s", DEFAULT_DNS_GROUP)) end fs.symlink(TMP_CONF_FILE, SMARTDNS_CONF) sys.call(string.format('echo "conf-file %s" >> /etc/smartdns/custom.conf', string.gsub(SMARTDNS_CONF, appname, appname .. "*"))) log(" - 请让SmartDNS作为Dnsmasq的上游或重定向!")
294coder/Efficient-MIF
34,158
utils/log_utils.py
import os import json import logging from datetime import datetime from functools import partial import signal from typing import Any, Dict, List, Optional, Union, Protocol from collections.abc import Sequence, Iterable from contextlib import contextmanager import loguru import matplotlib.pyplot as plt import numpy as np import torch from torch.utils.tensorboard import SummaryWriter from torch import nn from torchvision.utils import make_grid import shortuuid from utils.misc import NameSpace, default, is_main_process from utils.visualize import get_spectral_image_ready import time from rich.console import Console from rich.logging import RichHandler from importlib import reload def get_time(sec): h = int(sec//3600) m = int((sec//60)%60) s = int(sec%60) return h,m,s class TimeFilter(logging.Filter): def filter(self, record): try: start = self.start except AttributeError: start = self.start = time.time() time_elapsed = get_time(time.time() - start) record.relative = "{0}:{1:02d}:{2:02d}".format(*time_elapsed) # self.last = record.relativeCreated/1000.0 return True def save2json_file(d: dict, path: str, mode: str = "w", indent: int = 4): assert path.endswith(".json"), "@path should end with .json" with open(path, mode) as f: json.dump(d, f, indent=indent) print(f"save json file in {path}") def loss_dict2str( loss_dict: "dict[int, float | torch.Tensor]", world_size: int = 1, round_fp: int = 6 ) -> str: log_str = "" for k, v in loss_dict.items(): log_str += f"{k}: {v / world_size:.{round_fp}f} " return log_str ## decrepted class TrainStatusLogger(object): def __init__( self, id="None", path="./train_status/status.pt", resume=False, args=None ): """ track training status as a context manager :param id: run id, which is defined in parser args :param path: pkl file's path :param resume: resume training. if you want to resume last training run, set id=None. if you want to specify one run to resume, pass a specified id. """ self.time_now = datetime.now() self.path = path self.id = id self.resume = resume self.status = { "id": id, "status": "running", "time_stamp": str(self.time_now.timestamp()), "time": self.time_now.strftime("%Y-%m-%d, %H:%M:%S"), "args": args, } self._base_path = os.path.dirname(self.path) if not os.path.exists(self.path): if not os.path.exists(self._base_path): os.mkdir(self._base_path) self.status_all = [self.status] else: self.status_all = self.load_train_status() self.check_unique_id() if resume: if id == "None": self.status, _ = self.find_last_untrained_status() else: self.status = self.find_run_by_id(id) # print('warning: argument @id is not equal to the resume id which will be ignored') else: self.status_all.append(self.status) # ======= handle the KeyboardInterrupt signal ======= def handler(*args): print('catch signal: KeyboardInterrupt') print('EXITTING...') raise KeyboardInterrupt signal.signal(signal.SIGINT, handler) @staticmethod def _check_status_legal(status): assert status in ("running", "done", "break") def load_train_status(self): if os.path.getsize(self.path) > 0: # with open(self.path, "rb") as f: # l = pickle.load(f) l = torch.load(self.path) else: raise EOFError("file is empty, you should delete it") print("load previous train status") return l def save_train_status(self): # with open(self.path, "wb") as f: # pickle.dump(self.status_all, f) torch.save(self.status_all, self.path) print("save all train status") def update_train_status(self, status): self._check_status_legal(status) self.status["status"] = status def find_last_untrained_status(self): f_sort = lambda d: d["time_stamp"] if d["status"] == "break" else "0" last_status = sorted(self.status_all, key=f_sort)[-1] return last_status, last_status["id"] def find_run_by_id(self, id): s = self._find_id(id) if s["status"] != "break": return s raise AttributeError( f"no id: {id} in not an existing run or has already been done" ) def _find_id(self, id): for s in self.status_all: if s["id"] == id: return s def print_status_by_id(self, id): s = self._find_id(id) for k, v in s: if isinstance(v, NameSpace): print(v) else: print(f"{k}: {v}") def check_unique_id(self): ids = [] for d in self.status_all: ids.append(d["id"]) assert len(ids) == len(np.unique(ids)), "exist id conflict" assert not self.status["id"] in ids or self.resume, ( "id conflicts, check your run id " "or delete all tracker pkl file. " f"the pkl file can be found in {self.path}" ) def __enter__(self): nbreak = 0 ndone = 0 for d in self.status_all: s = d["status"] if s == "done": ndone += 1 elif s == "break": nbreak += 1 print("=" * 20, "Log Train Process", "=" * 20, sep="") print(f"all runs: {ndone} run(s) done, {nbreak} run(s) break") return self def __exit__(self, exc_type, exc_val, exc_tb): # print(f'traceback: {exc_tb}') if exc_type is not None or exc_val is not None: print("=" * 20, "Find Error Happen", "=" * 20, sep="") print(f"catch error type: {exc_type}, error value: {exc_val}") self.update_train_status("break") else: print("=" * 20, "Training End", "=" * 20, sep="") self.update_train_status("done") # only save in main process if is_main_process(): self.save_train_status() def __repr__(self): def dict_str(d): s = "id: {:<10} status: {:<7} time_stamp: {:<20} time: {:<20}".format( d["id"], d["status"], d["time_stamp"], d["time"] ) return s repr = "" for d in self.status_all: repr += dict_str(d) + "\n" return repr def generate_id(length: int = 8) -> str: # ~3t run ids (36**8) run_gen = shortuuid.ShortUUID(alphabet=list("0123456789abcdefghijklmnopqrstuvwxyz")) return str(run_gen.random(length)) def easy_logger(level='INFO', format_str: str=None, func_name: str=None): reload(logging) if format_str is not None and func_name is not None: format_str = "%(func_name)s: " + format_str elif format_str is None and func_name is not None: format_str = "%(func_name)s: %(message)s" else: format_str = "%(message)s" file_name = os.environ.get('LOG_FILE', None) if file_name: console = Console(file=open(file_name, 'a+')) else: console = None class FuncNameFilter(logging.Filter): def filter(self, record): record.func_name = func_name return True rich_handler = RichHandler(console=console, show_path=False, level=level, markup=True) rich_handler.setFormatter(logging.Formatter(format_str, datefmt='%X')) logger = logging.getLogger(__name__) logger.setLevel(level) logger.addHandler(rich_handler) class ProtocalLogger(Protocol): @classmethod def print(ctx, *msg, sep=" ", level: str | int="INFO"): msg = sep.join(map(str, msg)) if isinstance(level, str): level = eval(f'logging.{level.upper()}') logger.log(level, msg, extra={"markup": True}) @classmethod def debug(ctx, msg): pass @classmethod def info(ctx, msg): pass @classmethod def warning(ctx, msg): pass @classmethod def error(ctx, *msg, sep=" ", raise_error: bool=False, error_type=None): msg = sep.join(map(str, msg)) ctx.print(msg, level='ERROR') if raise_error: if error_type is not None: raise error_type(msg) raise RuntimeError(msg) # add filter to add func_name to log if func_name is not None: logger.addFilter(FuncNameFilter()) # logger signature logger: ProtocalLogger # add attributes to logger logger.print = ProtocalLogger.print logger.debug = partial(ProtocalLogger.print, level='DEBUG') logger.info = partial(ProtocalLogger.print, level='INFO') logger.warning = partial(ProtocalLogger.print, level='WARNING') logger.error = ProtocalLogger.error logger._console = rich_handler.console return logger from loguru import logger, _logger class LoguruLogger: _logger = logger console = None handler = [] _first_import = True _default_file_format = "<green>[{time:MM-DD HH:mm:ss}]</green> <level>[{level}] {message}</level>" _default_console_format = "[{time:HH:mm:ss}] <level>[{level}] {message}</level>" @classmethod def logger(cls, sink=None, format=None, filter=None, **kwargs) -> "_logger.Logger": reload(loguru) if cls._first_import: cls._logger.remove() # the first time import cls.console = Console(color_system=None) cls._logger.add( default(sink, lambda x: cls.console.print(x)), colorize=True, format=default(format, cls._default_console_format), **kwargs ) cls._first_import = False else: if sink is not None: handler = cls._logger.add(sink, format=default(format, cls._default_file_format), filter=filter, **kwargs) cls.handler.append(handler) return cls._logger @classmethod def add(cls, *args, **kwargs): handler = cls._logger.add(*args, **kwargs) cls.handler.append(handler) @classmethod def remove_all(cls): for h in cls.handler: cls._logger.remove(h) cls.handler = [] @classmethod def remove_id(cls, id): cls._logger.remove(id) @classmethod def bind(cls, *args, **kwargs): return cls._logger.bind(*args, **kwargs) @contextmanager def catch_any_error(): try: logger = LoguruLogger.logger() yield logger except Exception as e: logger.error(f"catch error: {e}", raise_error=True) logger.exception(e) finally: LoguruLogger.remove_all() def get_logger( base_path: str = None, name: str = None, args=None, std_level=logging.INFO, file_level: Union[tuple, int] = (logging.DEBUG,), file_handler_names: Union[tuple, str] = ("debug",), file_mode: str = "w", show_pid: bool = False, method_dataset_as_prepos=True, ): """ get logger to log :param base_path: such like './log/' :param name: logger name such as 'train_epoch_300' :param std_level: stream level :param file_level: file level :param file_handler_names: :param file_mode: 'a' append, 'w' write :param show_pid: show thread id :return: logger and List[handlers] """ assert name is not None, "@param name should not be None" reload(logging) # assert base_path is not None, "@param base_path should not be None" if not show_pid: # format_str = "[%(asctime)s - %(funcName)s]-%(levelname)s: %(message)s" # format_str = "(%(relative)s) %(message)s" format_str = "%(message)s" else: # format_str = ( # "[%(asctime)s - %(funcName)s - pid: %(thread)d]-%(levelname)s: %(message)s" # ) format_str = "pid: %(thread)d) %(message)s" rich_handler = RichHandler(show_path=False, markup=True) rich_handler.setFormatter(logging.Formatter(format_str, datefmt='%X')) logger = logging.getLogger(name) logger.setLevel(logging.DEBUG) logger.addHandler(rich_handler) hdls = [] # stream_handler = logging.StreamHandler(sys.stdout) # stream_handler.setLevel(std_level) # hdls.append(stream_handler) # logger.addHandler(stream_handler) if base_path is not None: assert len(file_handler_names) == len(file_level), \ "@param file_handler_names and @param file_level should be list and equal length" for n, level in zip(file_handler_names, file_level): if method_dataset_as_prepos: file_log_dir = os.path.join(base_path, args.full_arch, args.dataset, name) else: file_log_dir = os.path.join(base_path, name) if not os.path.exists(file_log_dir): os.makedirs(file_log_dir) print(f"logging: make log file [{os.path.abspath(file_log_dir)}]") file_log_path = os.path.join(file_log_dir, n + ".log") # file_handler = logging.FileHandler(file_log_path, mode=file_mode) file_console = Console(file=open(file_log_path, 'w')) file_handler = RichHandler(console=file_console, show_path=False, markup=True) # formatter = logging.Formatter( # "[%(asctime)s - %(name)s] - %(levelname)s: %(message)s" # ) # file_handler.setFormatter(formatter) file_handler.setLevel(level) hdls.append(file_handler) logger.addHandler(file_handler) else: file_console = None file_handler = None file_log_dir = None file_log_path = None for handler in logger.handlers: handler.addFilter(TimeFilter()) def log_print(*msg, sep=" ", level="INFO", dist=False, proc_id=None): if dist or is_main_process(): if isinstance(level, str): level_int = eval(f"logging.{level}") msgs = f"{proc_id=} - " if proc_id is not None else "" msgs += sep.join(map(str, msg)) logger.log(level=level_int, msg=msgs) # register some attrs logger.print = log_print logger._console = rich_handler.console logger._file_console = file_console logger._file_path = file_log_path return logger, hdls, file_log_dir # class WandbLogger: # def __init__( # self, # project_name, # run_name=None, # save_path=None, # config=None, # resume=None, # **kwargs, # ): # self.run = wandb.init( # project=project_name, # name=run_name, # dir=save_path, # config=config, # resume=resume, # settings=wandb.Settings(start_method="fork"), # **kwargs, # ) # def watch(self, type, network, graph=False, freq=20): # # type(str) One of "gradients", "parameters", "all", or None # wandb.watch(network, log=type, log_graph=graph, log_freq=freq) # def log_image(self, tensor_or_numpy, img_name, *args): # # e.g. cols = ['id','pred', 'gt', 'res'] # # tensor_or_numpy should be tensor [B, C, H, W] or numpy array [H, W, C] # # C can only be 1, 3 or 4 # # warning: you should control B for clear visualization, suggest 2<B<7 # if isinstance(tensor_or_numpy, torch.Tensor): # x = tensor_or_numpy.cpu().numpy() # elif isinstance(tensor_or_numpy, np.ndarray): # x = tensor_or_numpy # else: # raise NotImplementedError # assert x.shape[-1] in [1, 3, 4] # img_log = [wandb.Image(xi) for xi in x] if x.ndim == 4 else [wandb.Image(x)] # self.run.log({img_name: img_log}) # def log_table(self, table_data, cols, table_name, *args): # # for example # # my_data = [ # # [0, wandb.Image("img_0.jpg"), 0, 0], # # [1, wandb.Image("img_1.jpg"), 8, 0], # # [2, wandb.Image("img_2.jpg"), 7, 1], # # [3, wandb.Image("img_3.jpg"), 1, 1] # # ] # # # # create a wandb.Table() with corresponding columns # # columns =["id", "image", "prediction", "truth"] # # test_table = wandb.Table(data=my_data, columns=columns) # # add_data # # add row: table.add_data("3a", "3b", "3c") # # add col: table.add_column(name="col_name", data=col_data) # wandb.log({table_name: wandb.Table(data=table_data, columns=cols)}) # def log_curve(self, value=None, name=None, d=None, *args): # if d is None: # wandb.log({name: value}) # else: # wandb.log(d) class NoneLogger: def __init__(self, *args, **kwargs): class NoneWriter: def __init__(self) -> None: pass def close(self, *args, **kwargs): pass self.writer = NoneWriter() self.logger, *_ = get_logger(name=kwargs.pop('name')) # add time and run_id args = kwargs['cfg'] args.logger_config.name = ( time.strftime("%Y-%m-%d-%H-%M-%S", time.localtime()) + "_" + args.logger_config.name ) args.logger_config.name += "_" + args.run_id + f"_{args.comment}" name = args.logger_config.name self.log_file_dir = os.path.join(args.logger_config.base_path, args.full_arch, args.dataset, name) @property def console(self): return self.logger._console @property def file_console(self): return self.file_console._file_console def watch(self, *args, **kwargs): pass def log_image(self, *args, **kwargs): pass def log_images(self, *args, **kwargs): pass def log_curve(self, *args, **kwargs): pass def log_curves(self, *args, **kwargs): pass def log_network(self, *args, **kwargs): pass def print(self, *msg, level="INFO", dist=False, proc_id=None): if dist or is_main_process(): if isinstance(level, str): level_int = eval(f"logging.{level}") msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.logger.log(level=level_int, msg=msgs) def info(self, *msg, dist=False, proc_id=None): if dist or is_main_process(): level_int = logging.INFO msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.logger.log(level=level_int, msg=msgs) def debug(self, *msg, dist=False, proc_id=None): if dist or is_main_process(): level_int = logging.DEBUG msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.logger.log(level=level_int, msg=msgs) def warning(self, *msg, dist=False, proc_id=None): if dist or is_main_process(): level_int = logging.WARNING msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.logger.log(level=level_int, msg=msgs) def error(self, *msg, dist=False, proc_id=None): if dist or is_main_process(): level_int = logging.ERROR msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.logger.log(level=level_int, msg=msgs) class TensorboardLogger: def __init__( self, args=None, tsb_logdir=None, comment=None, file_stream_log=True, config_file_mv="./configs", config_file_type="yaml", method_dataset_as_prepos=False ): """ Args: args: config args from main.py tsb_logdir: tensorboard log dir comment: file_stream_log: file stream dir config_file_mv: where arch_config.yaml dir at """ self.grad_dict = {} # if not os.path.exists(logdir): # os.mkdir(logdir) self.hooks = {} self.watch_type = "None" self.freq = 10 # add time and run_id args.logger_config.name = ( time.strftime("%Y-%m-%d-%H-%M-%S", time.localtime()) + "_" + args.logger_config.name ) args.logger_config.name += "_" + args.run_id + f"_{args.comment}" if file_stream_log: self.file_logger, self.file_hdls, self.log_file_dir = get_logger( **args.logger_config.to_dict(), args=args, method_dataset_as_prepos=method_dataset_as_prepos ) config_cp_path = os.path.join(self.log_file_dir, "config.json") save2json_file(args.to_dict(), config_cp_path) # shutil.copy2(os.path.join(config_file_mv, f'{args.arch}_config.{config_file_type}'), self.log_file_dir) self.print( f"\nmove config file to {os.path.abspath(self.log_file_dir)}", level="INFO" ) self.writer = SummaryWriter(default(tsb_logdir, self.log_file_dir), comment) # set file path to os.environ os.environ['LOG_FILE'] = self.logger_file_path self.file_logger.debug(f'set os environ `LOG_FILE`={self.logger_file_path}') @property def console(self) -> Console: return self.file_logger._console @property def logger_file_path(self) -> str: return self.file_logger._file_path @property def file_console(self): return self.file_console._file_console def check_tensor_float(self, x): if isinstance(x, torch.Tensor): if x.dtype != torch.float32: x = x.to(dtype=torch.float32) return x @is_main_process def watch(self, network: nn.Module, watch_type: str, freq: int): assert watch_type in ( "all", "grad", "None", ), "@watch_type should only be all, grad or None" if watch_type == "None": return self.watch_type = watch_type self.freq = freq def _hook(grad, name): self.grad_dict[name] = grad for n, p in network.named_parameters(): hook = partial(_hook, name=n) self.hooks[n] = hook p.register_hook(hook) @is_main_process def log_curve(self, x, name, epoch): self.writer.add_scalar(name, self.check_tensor_float(x), epoch) @is_main_process def log_curves(self, x_dict: Dict, epoch): # for example: # for i in range(100): # writer.add_scalars('run_14h', {'xsinx': i * np.sin(i / r), # 'xcosx': i * np.cos(i / r), # 'tanx': np.tan(i / r)}, i) # self.writer.add_scalars(main_name, x_dict, epoch) for k, v in x_dict.items(): self.writer.add_scalar(k, self.check_tensor_float(v), epoch) @is_main_process def log_image(self, image, name, epoch): if image.ndim == 3: assert image.shape[0] <= 3, ( f"the number of image channel " f"should not greater than 3 but got shape {image.shape}" ) self.writer.add_image(name, self.check_tensor_float(image), epoch, dataformats="CHW") @is_main_process def log_images(self, batch_imgs: Sequence, nrow: int, names: Sequence, task: str, epoch: int, ds_name: str, **grid_kwargs): assert task in ['fusion', 'sharpening'], '@task should be fusion or sharpening' for batch_img, name in zip(batch_imgs, names): batch_img = get_spectral_image_ready(self.check_tensor_float(batch_img), name, task, ds_name) grid_img = make_grid(batch_img, nrow=nrow, **grid_kwargs) self.log_image(grid_img, name, epoch) @is_main_process def log_network(self, network: nn.Module, ep: int): if self.watch_type != "None": if ep % self.freq == 0: for (_, g), (n, p) in zip( self.grad_dict.items(), network.named_parameters() ): if self.watch_type == "all": self.writer.add_histogram(n + "_data", p.flatten().float(), ep) self.writer.add_histogram(n + "_grad", g.flatten().float(), ep) def print(self, *msg, level="INFO", dist=False, proc_id=None): if dist or is_main_process(): if isinstance(level, str): level_int = eval(f"logging.{level}") msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.file_logger.log(level=level_int, msg=msgs) def info(self, *msg, dist=False, proc_id=None): if dist or is_main_process(): level_int = logging.INFO msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.file_logger.log(level=level_int, msg=msgs) def debug(self, *msg, dist=False, proc_id=None): if dist or is_main_process(): level_int = logging.DEBUG msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.file_logger.log(level=level_int, msg=msgs) def warning(self, *msg, dist=False, proc_id=None): if dist or is_main_process(): level_int = logging.WARNING msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.file_logger.log(level=level_int, msg=msgs) def error(self, *msg, dist=False, proc_id=None, raise_error: bool=False, error_type=None): if dist or is_main_process(): level_int = logging.ERROR msgs = f"{proc_id=} - " if proc_id is not None else "" for s in msg: msgs += s self.file_logger.log(level=level_int, msg=msgs) if raise_error: if error_type is not None: raise error_type(msgs) raise RuntimeError(msgs) # from aim import Distribution, Image, Repo, Run, Session, Text # class AimLogger(object): # def __init__( # self, run_name, resume_hash_name=None, hparams=None, cap_term_logs=True, # *, # desp=None, # ): # """Aim framewarke logger # Args: # run_name (str): experiment name # resume_hash_name (str, optional): set it if you want to resume training. Defaults to None. # hparams (dict, optional): dict of some hyperparameters of your experiments. Defaults to None. # cap_term_logs (bool, optional): capture terminal logs. Defaults to True. # desp (str, optional): short description of your experiment. Defaults to None. # """ # self.run = Run( # resume_hash_name, # experiment=run_name, # capture_terminal_logs=cap_term_logs, # log_system_params=False, # ) # setattr(self.run, 'description', desp) # self.run["hparams"] = hparams # self.log_info("log params:", hparams) # self.run_name = run_name # self.repo = Repo("./") # def _convert_uint8_img(self, img): # # TODO: to suit PIL package, convert the image into unit8 type # # ref to https://pillow.readthedocs.io/en/stable/handbook/concepts.html # # Aim package only support uint8 type image :> # if isinstance(img, torch.Tensor): # img = img.detach().cpu().numpy() # img = img - img.min() # img = img / img.max() # img = (img * 255).astype(np.uint8) # return img # def log_image(self, image, name=None, epoch=None, context=None): # # check image type # if isinstance(image, np.ndarray): # assert image.ndim in [3, 2] # img = Image(self._convert_uint8_img(image)) # # TODO: convert the Tensor into ndarray is really slow # elif isinstance(image, torch.Tensor): # assert image.ndim in [4, 3, 2] # if image.ndim == 4: # nrows = math.sqrt(image.shape[0]) # image = make_grid(image, nrow=nrows) # img = Image(image) # elif image.ndim in [3, 2]: # img = Image(self._convert_uint8_img(image)) # else: # self.run.log_warning(f"not support image shape {image.shape}") # elif isinstance(image, plt.Figure): # img = Image(image) # else: # img = image # self.run.track(img, name=name, epoch=epoch, context=context) # def log_text(self, text, name=None, epoch=None, context=None): # self.run.track(Text(text), name=name, epoch=epoch, context=context) # @beartype.beartype() # def log_metrics(self, metrics, name=None, epoch=None, context: dict = None): # """log metrics or other values # Args: # metrics (dict or values): a dict or values to log # epoch (_type_): _description_ # context (_type_, optional): _description_. Defaults to None. # """ # self.run.track(metrics, name=name, epoch=epoch, context=context) # @beartype.beartype() # def log_distribution( # self, # distribution, # name=None, # epoch=None, # context: dict = None, # ): # ####################################################################### # # !!!! # # warning: the context must be a dict or it will explode your aim repo # # I don't know why, maybe it is a bug # ####################################################################### # distribution = distribution.flatten() # if isinstance(distribution, torch.Tensor): # distribution = distribution.detach().cpu().numpy() # # hist, bins = np.histogram(distribution, bins=64 if 64 < len(distribution) else len(distribution)) # # bin range is # # bin_range = [bins[0], bins[-1]] # self.run.track( # Distribution(distribution), name=name, epoch=epoch, context=context # ) # def log_network(self, network: nn.Module, epoch: int = None): # # refer to the warning in @log_distribution, it's important # # I set the @context into a dict, do not change it # for n, p in network.named_parameters(): # p = p.flatten().detach().cpu().numpy() # # context = {"net_params_dist": n} # self.log_distribution( # p, name="network_params", epoch=epoch, context={"net_params_dist": n} # do not change # ) # def close(self): # # close the run # # finilize and close, may one of them will take effect # self.run.finalize() # self.run.close() # print("logger closed") # def _make_msg_one_text(self, *msg): # if isinstance(msg[0], str) and len(msg) == 1: # return msg[0] # fin_msg = "" # for m in msg: # fin_msg += str(m) # return fin_msg # ######## override those functions ######### # def log_info(self, *msg): # self.run.log_info(self._make_msg_one_text(*msg)) # def log_warning(self, *msg): # self.run.log_warning(self._make_msg_one_text(*msg)) # def log_error(self, *msg): # self.run.log_error(self._make_msg_one_text(*msg)) # def log_debug(self, *msg): # self.run.log_debug(self._make_msg_one_text(*msg)) # ########################################### # # repo control # # may be not used # def delete_run(self, run_hash=None): # # sometimes the run is locked, so we need to release it # # or just delete it # print( # "warning, deleting run: {}".format(run_hash if run_hash else self.run.hash) # ) # run = self.repo.get_run(run_hash if run_hash else self.run.hash) # # run._lock.release() # will raise Nonetype do not have the attribute # run.read_only = False # ans = input("press [y/n] to confirm deleting") # if ans == "y": # d_m = self.repo.delete_run(run_hash) # print("deleted {}: {}".format(run.hash, d_m)) # elif ans == "n": # print("canceled") # else: # print("invalid input, canceled") # @property # def hash(self): # return self.run.hash if __name__ == '__main__': logger = easy_logger() logger.info('[dim green] 1234 [/dim green]')
294coder/Efficient-MIF
10,992
utils/visualize.py
import os import os.path as osp from typing import Literal, Tuple, Union import cv2 import numpy as np import torch from torch import Tensor import matplotlib.pyplot as plt from utils.misc import to_numpy def get_rgb_channel_by_dataset_name(tensor, dataset_name: str): if dataset_name in ('wv3', 'wv2'): return tensor[:, [4,2,0], ...] elif dataset_name in ('gf2', 'qb'): return tensor[:, :3, ...] elif dataset_name in ('gf5', 'gf5-gf1'): return tensor[:, [40, 30, 20], ...] elif dataset_name == 'houston': return tensor[:, [39, 29, 19], ...] elif 'cave' in dataset_name or 'harvard' in dataset_name: return tensor[:, [29, 19, 9], ...] else: return tensor[:, :3, ...] def permute_dim(*args): d = [ i.permute(2, 0, 1) for i in args ] return d def normalize(img, to_uint8=True): """ centering image to show :param img: numpy array, shape [H, W, C] :return: uint8 type image """ img = img - img.min((0, 1)) img = img / img.max((0, 1)) if to_uint8: img *= 255 img = img.astype('uint8') return img def invert_normalized(norm_img: Union[Tensor, np.ndarray], mean: Union[Tensor, np.ndarray], std: Union[Tensor, np.ndarray], *, change_back_dim=True): """ invert image normalized to unnormalized image Args: norm_img: Tensor: [C, H, W] or [B, C, H, W] mean: Tensor, [C, ] std: Tensor, [C, ] change_back_dim: bool, change channel dim back Returns: Ndarray, unnormalized image """ if isinstance(norm_img, Tensor): norm_img = to_numpy(norm_img)[0] if isinstance(mean, Tensor): mean = to_numpy(mean)[0] if isinstance(std, Tensor): std = to_numpy(std)[0] if norm_img.ndim == 4: _dim_trans = [0, 2, 3, 1] _dim_back_trans = [0, -1, 1, 2] else: _dim_trans = [1, 2, 0] _dim_back_trans = [2, 0, 1] norm_img = norm_img.transpose(_dim_trans) assert norm_img.shape[-1] == mean.size == std.size unnormed_img = norm_img * std + mean # [H, W, C] or [B, H, W, C] if change_back_dim: unnormed_img = unnormed_img.transpose(_dim_back_trans) return unnormed_img def hist_equal(img): """ equalize an image :param img: numpy array, shape [H, W, C], C can be any int :return: """ if img.ndim == 3: for i in range(img.shape[-1]): img[..., i] = cv2.equalizeHist(img[..., i]) else: img = cv2.equalizeHist(img) return img def res_image(gt: Tensor, sr: Tensor, *, exaggerate_ratio: int = None) -> torch.Tensor: # shape [B, C, H, W] ratio = exaggerate_ratio if exaggerate_ratio is not None else 1. res = torch.abs(gt - sr).mean(1, keepdim=True) * ratio return res def get_spectral_image_ready(batch_image: Tensor, tensor_name: str, task: str=None, ds_name: Literal['wv3', 'wv2', 'gf2', 'qb', 'gf5', 'gf5-gf1', 'flir', 'tno', 'msrs']=None) -> Tensor: # batch_image: [B, C, H, W] img_arrs = batch_image.permute(0, 2, 3, 1).detach().cpu().numpy() # [B, H, W, C] if task == 'fusion': transform_fn = lambda x, const: torch.tensor(x, dtype=torch.float32) elif task == 'sharpening': if tensor_name in ('lms', 'ms', 'sr'): # for pansharpening and HISR tasks batch_image = get_rgb_channel_by_dataset_name(batch_image, ds_name) elif tensor_name == 'pan' and batch_image.shape[1] > 3: batch_image = batch_image[:, :3] transform_fn = lambda x, const: torch.tensor(normalize(x, to_uint8=False) * const) else: raise ValueError(f'Invalid task: {task}') if 'res' in tensor_name: equalized_img = [transform_fn(i, 10).permute(-1, 0, 1)[None, ...] for i in img_arrs] # [1, C, H, W] else: equalized_img = [transform_fn(i, 1).permute(-1, 0, 1)[None, ...] for i in img_arrs] grid = torch.cat(equalized_img, dim=0) return grid def viz_batch(img: Tensor, base_path='./visualized_img', suffix=None, start_index=1, format='jpg'): assert suffix is not None, 'arg @suffix can not be None' assert suffix in ['pan', 'ms', 'sr', 'gt', 'residual'], 'arg @suffix should only be pan, ms or sr' img_arrs = img.permute(0, 2, 3, 1).numpy() if suffix == 'residual': equalized_img = [i for i in img_arrs] else: # equalized_img = [hist_equal(normalize(i)) for i in img_arrs] equalized_img = [normalize(i) for i in img_arrs] path = osp.join(base_path, suffix) if not osp.exists(path): os.makedirs(path) # all_path = path + '.mat' # savemat(all_path, {f'{suffix}': img_arrs}) # fig, ax = plt.subplots(figsize=(5, 5), dpi=100) # ax: plt.Axes # fig: plt.Figure for i, img in enumerate(equalized_img, start_index): h, w = img.shape[-2:] img_path1 = osp.join(path, str(i) + '.' + format) # plt.cla() if suffix == 'pan' or suffix == 'residual': # ax.imshow(img, cmap='gray') cv2.imwrite(img_path1, img) # elif suffix == 'residual': # ax.imshow(img) else: try: # ax.imshow(img[..., [0, 2, 4]]) cv2.imwrite(img_path1, img[..., [0, 2, 4]]) except: # ax.imshow(img[..., :3]) cv2.imwrite(img_path1, img[..., :3]) # ax.set_axis_off() # fig.set_size_inches(h, w) # fig.savefig(img_path1, format=format, dpi=50, bbox_inches='tight', pad_inches=0.) # plt.close() def show_details(img: np.ndarray, cpos_ratio: Tuple[float, float], area_pixels: Tuple[int, int], interp_ratio: int = 3, color: Tuple[int, int, int] = (0, 255, 0), thickness: int = 2, place: str = None) -> np.ndarray: """select a patch in raw image which decided by @cpos_ratio and @area_pixels, the function will interpolate @interp_ratio times and paste it in a corner. Args: img (np.ndarray): raw image needed to detailed, format [H, W, C] cpos_ratio (Tuple[float, float]): selected patch's centroid, from 0 to 1 area_pixels (Tuple[int, int]): pixel area of the patch interp_ratio (int, optional): interpolate ratio. Defaults to 3. color (Tuple[int, int, int], optional): color of the box. Defaults to (0, 255, 0). recommend colors: (236,229,240) (233,138,21) (0,59,54) thickness (int, optional): thickness of the box. Defaults to 2. place(str, optional): where to place the interpolated patch. """ assert 0 < cpos_ratio[0] < 1 and 0 < cpos_ratio[1] < 1, '@cpos_ratio can only be range (0, 1)' # to array if img.ndim == 2: img = np.repeat(img[..., np.newaxis], 3, axis=-1) elif img.shape[-1] == 1: img = np.repeat(img, 3, axis=-1) img_size = np.array(img.shape[:2]) cpos_pixels = np.array(cpos_ratio) * img_size cpos_ratio = np.array(cpos_ratio) area_pixels = np.array(area_pixels) paste_pixels = area_pixels * interp_ratio img = img.astype('uint8') img = np.ascontiguousarray(img) # bound check bound = [] for i, j in zip((-1, 1), (-1, 1)): bound.append([i * area_pixels[0] / 2, j * area_pixels[1] / 2]) bound = np.array(bound) bound = np.repeat(cpos_pixels[np.newaxis, ...], 2, axis=0) + bound assert not np.bitwise_or(bound[0] < 0, bound[1] > img_size).any(), \ f'selected range out of image size, image size {img_size} but get selected range {bound}' # find furthest corner to paste the interpolated patch if place is None: furthest_pos_ratio = None furthest_dis = 0. for i in (0, 1): for j in (0, 1): d = (i - cpos_ratio[0]) ** 2 + (j - cpos_ratio[1]) ** 2 if d > furthest_dis: furthest_pos_ratio = (i, j) furthest_dis = d else: assert place in ('lt', 'rt', 'lb', 'rb'), '@place should be one of [lt, rt, lb, rb]' place_dict = {'lt': (0, 0), 'rt': (0, 1), 'lb': (1, 0), 'rb': (1, 1)} furthest_pos_ratio = place_dict[place] bound = bound.astype('int') patch = img[bound[0, 0]:bound[1, 0], bound[0, 1]: bound[1, 1], :] interp_img = cv2.resize(patch, dsize=paste_pixels[::-1]) box_edge_point = [] cv2.rectangle(img, bound[0][::-1], bound[1][::-1], color, thickness=thickness) box_pre_thick = thickness // 2 if furthest_pos_ratio == (0, 0): img[:paste_pixels[0], :paste_pixels[1], :] = interp_img box_edge_point = [[box_pre_thick, box_pre_thick], paste_pixels[::-1]] elif furthest_pos_ratio == (1, 0): img[-paste_pixels[0]:, :paste_pixels[1], :] = interp_img box_edge_point = [[box_pre_thick, img_size[0] - paste_pixels[0]], [paste_pixels[1], img_size[0] - box_pre_thick]] elif furthest_pos_ratio == (0, 1): img[:paste_pixels[0], -paste_pixels[1]:, :] = interp_img box_edge_point = [[img_size[1] - paste_pixels[1], box_pre_thick], [img_size[1] - box_pre_thick, paste_pixels[0]]] else: img[-paste_pixels[0]:, -paste_pixels[1]:, :] = interp_img box_edge_point = [[img_size[1] - paste_pixels[1], img_size[0] - paste_pixels[0]], [img_size[1] - box_pre_thick, img_size[0] - box_pre_thick]] cv2.rectangle(img, box_edge_point[0], box_edge_point[1], color, thickness) return img def plt_plot_img_without_white_margin(img, *args, **kwargs): """ :param img: format [H, W, C] :param args: plt.imshow args :param kwargs: plt.imshow kwargs :return: """ width, height = img.shape[:2] ax = plt.imshow(img, *args, **kwargs) fig = plt.gcf() fig.set_size_inches(width / 100, height / 100) plt.gca().xaxis.set_major_locator(plt.NullLocator()) plt.gca().yaxis.set_major_locator(plt.NullLocator()) plt.subplots_adjust(top=1, bottom=0, left=0, right=1, hspace=0, wspace=0) plt.margins(0, 0) plt.gca().set_axis_off() return fig, ax if __name__ == '__main__': import matplotlib.pyplot as plt import PIL.Image as Image img = Image.open('../visualized_img/sr/8.eps') img = np.asarray(img) img = show_details(img, cpos_ratio=(0.2, 0.8), area_pixels=(50, 50), thickness=2) plt_plot_img_without_white_margin(img) plt.show()
2977094657/DidaAPI
1,685
frontend/docs/api/statistics/general-statistics.md
# 获取通用统计信息 ## 接口信息 - **接口路径**: `GET /statistics/general` - **接口描述**: 获取概览、成就值、趋势等通用统计信息 - **请求方式**: GET - **认证要求**: 需要先完成微信登录获取认证会话 ## 请求参数 无需参数 ## 请求示例 ```bash curl -X GET "http://localhost:8000/statistics/general" ``` ## 响应格式 ### 成功响应 ```json { "score": 1250, "level": 5, "yesterdayCompletedCount": 8, "todayCompletedCount": 12, "totalCompletedCount": 1580, "todayPomoCount": 6, "yesterdayPomoCount": 4, "totalPomoCount": 320, "todayFocusTime": 150, "yesterdayFocusTime": 100, "totalFocusTime": 8000, "targetPomoCount": 8, "targetFocusTime": 200 } ``` ### 错误响应 #### 未认证 ```json { "error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录" } ``` #### 服务错误 ```json { "error": "service_error", "message": "获取通用统计信息失败,请稍后重试" } ``` ## 响应字段说明 | 字段名 | 类型 | 描述 | |--------|------|------| | score | integer | 成就值 | | level | integer | 账号等级 | | yesterdayCompletedCount | integer | 昨日完成任务数 | | todayCompletedCount | integer | 今日完成任务数 | | totalCompletedCount | integer | 总完成任务数 | | todayPomoCount | integer | 今日番茄数 | | yesterdayPomoCount | integer | 昨日番茄数 | | totalPomoCount | integer | 总番茄数 | | todayFocusTime | integer | 今日专注时长(分钟) | | yesterdayFocusTime | integer | 昨日专注时长(分钟) | | totalFocusTime | integer | 总专注时长(分钟) | | targetPomoCount | integer | 目标番茄数 | | targetFocusTime | integer | 目标专注时长(分钟) | ## 使用说明 1. **认证要求**: 需要先调用微信登录接口获取认证会话 2. **统计范围**: 包含任务完成和番茄专注的综合统计 3. **时间维度**: 提供昨日、今日和总计三个维度的数据 4. **目标对比**: 包含目标设置,便于进度对比 ## 相关接口 - [获取用户排名统计](../statistics.md) - [获取任务统计信息](./task-statistics.md) - [获取番茄专注概览](../pomodoros.md) - [微信登录流程](../auth/wechat-login-flow.md) ## 注意事项 - 需要先完成微信登录获取认证会话 - 时间以分钟为单位 - 数据实时更新 - 成就值和等级基于用户活跃度计算
2977094657/DidaAPI
2,950
frontend/docs/api/tasks/get-trash-tasks.md
# 获取垃圾桶任务 获取垃圾桶中的任务列表。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/project/all/trash/page` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 ### Query Parameters | 参数名 | 类型 | 必填 | 说明 | 示例 | |--------|------|------|------|------| | limit | number | 否 | 每页任务数量,默认50 | 50 | | task_type | number | 否 | 任务类型,默认1 | 1 | ## 响应格式 ### 成功响应 返回垃圾桶任务的原始数据: ```json { "tasks": [ { "id": "string", "projectId": "string", "sortOrder": 0, "title": "string", "startDate": "string", "dueDate": "string", "timeZone": "string", "isFloating": true, "isAllDay": true, "reminders": [ { "id": "string", "trigger": "string" } ], "exDate": [ "string" ], "priority": 0, "status": 0, "items": [ "string" ], "modifiedTime": "string", "etag": "string", "deleted": 0, "createdTime": "string", "creator": 0, "attachments": [ "string" ], "columnId": "string", "parentId": "string", "childIds": [ "string" ], "kind": "string", "pinnedTime": "string", "imgMode": 0, "deletedBy": 0, "deletedTime": 0, "repeatFirstDate": "string", "repeatTaskId": "string", "repeatFrom": "string", "tags": [ "string" ], "commentCount": 0, "focusSummaries": [ { "userId": 0, "pomoCount": 0, "estimatedPomo": 0, "estimatedDuration": 0, "pomoDuration": 0, "stopwatchDuration": 0 } ], "repeatFlag": "string", "pomodoroSummaries": [ { "userId": 0, "count": 0, "estimatedPomo": 0, "duration": 0 } ] } ], "next": 0 } ``` ## 响应字段说明 ### 响应结构 | 字段名 | 类型 | 说明 | |--------|------|------| | tasks | array | 垃圾桶任务列表 | | next | number | 下一页标识 | ### 任务对象核心字段 | 字段名 | 类型 | 说明 | |--------|------|------| | id | string | 任务唯一标识符 | | projectId | string | 所属项目ID | | title | string | 任务标题 | | status | number | 任务状态 | | priority | number | 优先级 | | deleted | number | 删除标志 | | deletedBy | number | 删除者ID | | deletedTime | number | 删除时间 | ### 时间字段 | 字段名 | 类型 | 说明 | |--------|------|------| | createdTime | string | 创建时间 | | modifiedTime | string | 修改时间 | | startDate | string | 开始日期 | | dueDate | string | 截止日期 |
2977094657/DidaAPI
4,008
frontend/docs/api/tasks/get-completed-tasks.md
# 获取已完成/已放弃任务 获取用户的已完成或已放弃任务列表,支持分页获取。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/project/all/closed` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 ### Query Parameters | 参数名 | 类型 | 必填 | 说明 | 示例 | |--------|------|------|------|------| | to | string | 否 | 分页参数,使用上次响应最后一个任务的completedTime | "2025-03-15T13:30:54.000+0000" | | status | string | 否 | 任务状态,默认Completed | "Completed" 或 "Abandoned" | ### 状态参数说明 | 状态值 | 说明 | |--------|------| | Completed | 已完成的任务(默认) | | Abandoned | 已放弃的任务 | ### 分页机制 **重要说明**: 此接口的分页机制需要严格按照以下步骤执行,否则无法获取全部数据。 **第一次请求(不携带to字段)**: ``` GET https://api.dida365.com/api/v2/project/all/closed?from=&status=Completed ``` **获取已放弃任务(第一次请求)**: ``` GET https://api.dida365.com/api/v2/project/all/closed?from=&status=Abandoned ``` **后续分页请求**: ``` GET https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Completed GET https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Abandoned ``` **分页流程详解**: 1. **第一次请求**: 不携带`to`字段,直接请求 2. **获取分页参数**: 从响应中取最后一个任务的`completedTime`字段 3. **构造下次请求**: 将`completedTime`作为`to`参数 4. **重复请求**: 直到返回的任务数量少于50个为止 **to参数说明**: - 使用上次响应最后一个任务的`completedTime`字段值(原始格式) - 原始格式:`2025-03-15T13:30:54.000+0000` - API转换格式:`2025-03-15 13:30:54` - URL编码后:`2025-03-15%2013:30:54` - **注意**: 传入原始的completedTime格式,API会自动转换为正确格式 **分页终止条件**: - 当返回的任务数量少于50个时,表示已获取到最后一页 - 当返回空数组时,表示没有更多数据 ## 响应格式 ### 成功响应 返回已完成任务列表的原始数据: ```json [ { "id": "string", "projectId": "string", "sortOrder": 0, "title": "string", "content": "string", "desc": "string", "timeZone": "string", "isFloating": true, "isAllDay": true, "reminder": "string", "reminders": [ { "id": "string", "trigger": "string" } ], "repeatFirstDate": "string", "exDate": [ "string" ], "completedTime": "string", "completedUserId": 0, "repeatTaskId": "string", "priority": 0, "status": 0, "items": [ "string" ], "progress": 0, "modifiedTime": "string", "etag": "string", "deleted": 0, "createdTime": "string", "creator": 0, "repeatFrom": "string", "attachments": [ "string" ], "commentCount": 0, "focusSummaries": [ { "userId": 0, "pomoCount": 0, "estimatedPomo": 0, "estimatedDuration": 0, "pomoDuration": 0, "stopwatchDuration": 0, "focuses": [ [ "string" ] ] } ], "columnId": "string", "parentId": "string", "kind": "string", "pinnedTime": "string", "imgMode": 0, "startDate": "string", "dueDate": "string", "tags": [ "string" ], "pomodoroSummaries": [ { "userId": 0, "count": 0, "estimatedPomo": 0, "duration": 0 } ], "repeatFlag": "string", "childIds": [ "string" ] } ] ``` ## 响应字段说明 ### 核心字段 | 字段名 | 类型 | 说明 | |--------|------|------| | id | string | 任务唯一标识符 | | projectId | string | 所属项目ID | | title | string | 任务标题 | | content | string | 任务内容 | | completedTime | string | 完成时间(分页关键字段) | | priority | number | 优先级 | | status | number | 任务状态 | ### 时间字段 | 字段名 | 类型 | 说明 | |--------|------|------| | completedTime | string | 完成时间 | | createdTime | string | 创建时间 | | modifiedTime | string | 修改时间 | | startDate | string | 开始日期 | | dueDate | string | 截止日期 | ### 专注统计 | 字段名 | 类型 | 说明 | |--------|------|------| | focusSummaries | array | 专注统计摘要 | | pomodoroSummaries | array | 番茄钟统计摘要 |
2977094657/DidaAPI
11,083
frontend/docs/api/tasks/get-all-tasks.md
# 获取所有任务 获取当前用户的所有任务列表,包括任务详情、状态、优先级等信息。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/batch/check/0` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 认证要求 此接口需要完整的认证信息,包括: ### 必需的Cookie ```http Cookie: t={auth_token}; _csrf_token={csrf_token}; AWSALB={load_balancer_cookie}; AWSALBCORS={load_balancer_cookie} ``` ### 必需的请求头 ```http Accept: application/json, text/plain, */* Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7 Cache-Control: no-cache Origin: https://dida365.com Pragma: no-cache Referer: https://dida365.com/ User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36 X-Csrftoken: {csrf_token} X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":""} Hl: zh_CN X-Tz: Asia/Shanghai Traceid: {trace_id} ``` ## 完整请求示例 ```http GET https://api.dida365.com/api/v2/batch/check/0 HTTP/1.1 Host: api.dida365.com Accept: application/json, text/plain, */* Accept-Encoding: gzip, deflate, br, zstd Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7 Cache-Control: no-cache Cookie: t={auth_token}; _csrf_token={csrf_token} Hl: zh_CN Origin: https://dida365.com Pragma: no-cache Referer: https://dida365.com/ Traceid: {trace_id} User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36 X-Csrftoken: {csrf_token} X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":"{websocket_id}"} X-Tz: Asia/Shanghai ``` ## 响应格式 ### 成功响应 **状态码**: `200 OK` **Content-Type**: `application/json;charset=UTF-8` ### 响应体结构 ```json { "checkPoint": 0, "syncTaskBean": { "update": [ { "id": "string", "projectId": "string", "sortOrder": 0, "title": "string", "content": "string", "desc": "string", "timeZone": "string", "isFloating": true, "isAllDay": true, "reminder": "string", "reminders": [ { "id": "string", "trigger": "string" } ], "exDate": [ "string" ], "repeatTaskId": "string", "priority": 0, "status": 0, "items": [ "string" ], "progress": 0, "modifiedTime": "string", "etag": "string", "deleted": 0, "createdTime": "string", "creator": 0, "repeatFrom": "string", "attachments": [ { "id": "string", "refId": "string", "path": "string", "size": 0, "fileName": "string", "fileType": "string", "status": 0, "createdTime": "string" } ], "commentCount": 0, "focusSummaries": [ { "userId": 0, "pomoCount": 0, "estimatedPomo": 0, "estimatedDuration": 0, "pomoDuration": 0, "stopwatchDuration": 0, "focuses": [ [ "string" ] ] } ], "columnId": "string", "kind": "string", "imgMode": 0, "startDate": "string", "dueDate": "string", "repeatFlag": "string", "pomodoroSummaries": [ { "userId": 0, "count": 0, "estimatedPomo": 0, "duration": 0 } ], "childIds": [ "string" ], "pinnedTime": "string", "repeatFirstDate": "string", "tags": [ "string" ], "parentId": "string", "deletedBy": 0, "deletedTime": 0 } ], "tagUpdate": [ "string" ], "delete": [ "string" ], "add": [ "string" ], "empty": true }, "projectProfiles": [ { "id": "string", "name": "string", "isOwner": true, "color": "string", "sortOrder": 0, "sortOption": { "groupBy": "string", "orderBy": "string" }, "sortType": "string", "userCount": 0, "etag": "string", "modifiedTime": "string", "inAll": true, "showType": null, "muted": true, "reminderType": null, "closed": null, "transferred": null, "groupId": null, "viewMode": "string", "notificationOptions": [ "string" ], "teamId": null, "permission": "string", "kind": "string", "timeline": { "range": null, "sortType": "string", "sortOption": { "groupBy": "string", "orderBy": "string" } }, "needAudit": true, "barcodeNeedAudit": true, "openToTeam": true, "teamMemberPermission": null, "source": 0 } ], "projectGroups": [ "string" ], "filters": null, "tags": [ "string" ], "syncTaskOrderBean": { "taskOrderByDate": {}, "taskOrderByPriority": {}, "taskOrderByProject": {} }, "syncOrderBean": { "orderByType": { "taskPinned": { "{project_id_1}": { "changed": [ { "id": "string", "order": 0, "type": 0 } ], "deleted": [ { "id": "string", "order": 0, "type": 0 } ] }, "{project_id_2}": { "changed": [ { "id": "string", "order": 0, "type": 0 } ], "deleted": [ { "id": "string", "order": 0, "type": 0 } ] } } } }, "syncOrderBeanV3": { "orderByType": {} }, "inboxId": "string", "checks": null, "remindChanges": [ "string" ] } ``` ## 响应字段说明 ### 根级字段 | 字段名 | 类型 | 说明 | |--------|------|------| | checkPoint | number | 检查点 | | syncTaskBean | object | 任务同步数据 | | projectProfiles | array | 项目配置列表 | | projectGroups | array | 项目分组 | | filters | null | 过滤器 | | tags | array | 标签列表 | | syncTaskOrderBean | object | 任务排序数据 | | syncOrderBean | object | 排序数据 | | syncOrderBeanV3 | object | 排序数据V3 | | inboxId | string | 收件箱ID | | checks | null | 检查项 | | remindChanges | array | 提醒变更 | ### 任务字段 (syncTaskBean.update) | 字段名 | 类型 | 说明 | |--------|------|------| | id | string | 任务唯一标识符 | | projectId | string | 所属项目ID | | sortOrder | number | 排序顺序 | | title | string | 任务标题 | | content | string | 任务内容 | | desc | string | 任务描述 | | timeZone | string | 时区 | | isFloating | boolean | 是否浮动时间 | | isAllDay | boolean | 是否全天任务 | | reminder | string | 提醒设置 | | reminders | array | 提醒列表 | | exDate | array | 排除日期 | | repeatTaskId | string | 重复任务ID | | priority | number | 优先级 | | status | number | 任务状态 | | items | array | 子项目 | | progress | number | 完成进度 | | modifiedTime | string | 修改时间 | | etag | string | 实体标签 | | deleted | number | 删除状态 | | createdTime | string | 创建时间 | | creator | number | 创建者ID | | repeatFrom | string | 重复来源 | | attachments | array | 附件列表 | | commentCount | number | 评论数量 | | focusSummaries | array | 专注摘要 | | columnId | string | 列ID | | kind | string | 类型 | | imgMode | number | 图片模式 | | startDate | string | 开始日期 | | dueDate | string | 截止日期 | | repeatFlag | string | 重复标志 | | pomodoroSummaries | array | 番茄钟摘要 | | childIds | array | 子任务ID列表 | | pinnedTime | string | 置顶时间 | | repeatFirstDate | string | 首次重复日期 | | tags | array | 标签列表 | | parentId | string | 父任务ID | | deletedBy | number | 删除者ID | | deletedTime | number | 删除时间 | ### 项目配置字段 (projectProfiles) | 字段名 | 类型 | 说明 | |--------|------|------| | id | string | 项目ID | | name | string | 项目名称 | | isOwner | boolean | 是否为所有者 | | color | string | 项目颜色 | | sortOrder | number | 排序顺序 | | sortOption | object | 排序选项 | | sortType | string | 排序类型 | | userCount | number | 用户数量 | | etag | string | 实体标签 | | modifiedTime | string | 修改时间 | | inAll | boolean | 是否在全部中显示 | | showType | null | 显示类型 | | muted | boolean | 是否静音 | | reminderType | null | 提醒类型 | | closed | null | 是否关闭 | | transferred | null | 是否转移 | | groupId | null | 分组ID | | viewMode | string | 查看模式 | | notificationOptions | array | 通知选项 | | teamId | null | 团队ID | | permission | string | 权限 | | kind | string | 类型 | | timeline | object | 时间线配置 | | needAudit | boolean | 是否需要审核 | | barcodeNeedAudit | boolean | 条码是否需要审核 | | openToTeam | boolean | 是否对团队开放 | | teamMemberPermission | null | 团队成员权限 | | source | number | 来源 | ### 附件字段 (attachments) | 字段名 | 类型 | 说明 | |--------|------|------| | id | string | 附件ID | | refId | string | 引用ID | | path | string | 文件路径 | | size | number | 文件大小 | | fileName | string | 文件名 | | fileType | string | 文件类型 | | status | number | 状态 | | createdTime | string | 创建时间 | ### 专注摘要字段 (focusSummaries) | 字段名 | 类型 | 说明 | |--------|------|------| | userId | number | 用户ID | | pomoCount | number | 番茄钟数量 | | estimatedPomo | number | 预估番茄钟 | | estimatedDuration | number | 预估时长 | | pomoDuration | number | 番茄钟时长 | | stopwatchDuration | number | 秒表时长 | | focuses | array | 专注记录 | ## 相关接口 - [验证微信登录](../auth/validate-wechat-login.md) - 获取认证令牌 - [获取已完成任务](./get-completed-tasks.md) - 获取已完成任务列表 - [获取垃圾桶任务](./get-trash-tasks.md) - 获取垃圾桶任务列表
2977094657/DidaAPI
1,276
frontend/docs/api/tasks/get-tasks-summary.md
# 获取任务统计 ## 接口信息 - **接口路径**: `GET /tasks/summary` - **接口描述**: 获取任务的统计信息 - **请求方式**: GET - **认证要求**: 需要先完成微信登录获取认证会话 ## 请求参数 无需参数 ## 请求示例 ```bash curl -X GET "http://localhost:8000/tasks/summary" ``` ## 响应格式 ### 成功响应 ```json { "code": 200, "message": "获取任务统计成功", "data": { "total_tasks": 150, "completed_tasks": 120, "pending_tasks": 30, "completion_rate": 80.0 } } ``` ### 错误响应 #### 未认证 ```json { "error": "获取任务统计失败", "details": { "error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录" } } ``` #### 服务错误 ```json { "detail": "服务器内部错误: 具体错误信息" } ``` ## 响应字段说明 | 字段名 | 类型 | 描述 | |--------|------|------| | code | integer | 响应状态码,200表示成功 | | message | string | 响应消息 | | data | object | 统计数据对象 | | data.total_tasks | integer | 总任务数 | | data.completed_tasks | integer | 已完成任务数 | | data.pending_tasks | integer | 未完成任务数 | | data.completion_rate | float | 完成率(百分比) | ## 使用说明 1. **认证要求**: 需要先调用微信登录接口获取认证会话 2. **统计范围**: 统计当前用户的所有任务 3. **实时数据**: 返回实时的任务统计信息 4. **完成率计算**: 完成率 = (已完成任务数 / 总任务数) × 100 ## 相关接口 - [获取所有任务](./get-all-tasks.md) - [获取已完成任务](./get-completed-tasks.md) - [获取垃圾桶任务](./get-trash-tasks.md) - [微信登录流程](../auth/wechat-login-flow.md) ## 注意事项 - 需要先完成微信登录获取认证会话 - 统计数据基于用户的所有任务 - 完成率保留两位小数 - 如果没有任务,完成率为0
2977094657/DidaAPI
1,957
frontend/docs/api/auth/password-login.md
# 密码登录 使用用户名(邮箱或手机号)和密码进行滴答清单登录认证。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/user/signon?wc=true&remember=true` - **请求方法**: `POST` - **认证要求**: 无需认证 - **所属平台**: 滴答清单 ## 请求参数 ### 请求体 | 参数名 | 类型 | 必需 | 说明 | |----------|--------|------|--------------------------| | username | string | 是 | 登录账户(邮箱或手机号) | | password | string | 是 | 登录密码 | ### 请求示例 ```json { "username": "string", "password": "string" } ``` ## 响应格式 ### 成功响应 **HTTP状态码**: `200 OK` ```json { "token": "string", "userId": "string", "userCode": "string", "username": "string", "teamPro": false, "proStartDate": "string", "proEndDate": "string", "subscribeType": "string", "subscribeFreq": "string", "needSubscribe": false, "freq": "string", "inboxId": "string", "teamUser": false, "activeTeamUser": false, "freeTrial": false, "gracePeriod": false, "pro": true, "ds": false } ``` ## 响应字段说明 | 字段名 | 类型 | 说明 | |------------------|---------|------------------------| | token | string | 认证令牌,用于后续API调用 | | userId | string | 用户唯一ID | | userCode | string | 用户代码 | | username | string | 用户名(邮箱或手机号) | | teamPro | boolean | 是否为团队专业版 | | proStartDate | string | 专业版开始日期 | | proEndDate | string | 专业版结束日期 | | subscribeType | string | 订阅类型(如:wxpay_subscribe) | | subscribeFreq | string | 订阅频率(如:Month) | | needSubscribe | boolean | 是否需要订阅 | | freq | string | 频率 | | inboxId | string | 默认任务添加清单ID | | teamUser | boolean | 是否为团队用户 | | activeTeamUser | boolean | 是否为活跃团队用户 | | freeTrial | boolean | 是否为免费试用 | | gracePeriod | boolean | 是否在宽限期 | | pro | boolean | 是否为专业版 | | ds | boolean | 数据同步状态 |
281677160/openwrt-package
1,077
luci-app-passwall/root/usr/share/passwall/lease2hosts.sh
#!/bin/sh # dhcp.leases to hosts CONFIG=passwall TMP_PATH=/tmp/etc/${CONFIG} TMP_PATH2=/tmp/etc/${CONFIG}_tmp LOCK_FILE=/tmp/lock/${CONFIG}_lease2hosts.lock LEASE_FILE="/tmp/dhcp.leases" HOSTS_FILE="$TMP_PATH2/dhcp-hosts" TMP_FILE="/tmp/dhcp-hosts.tmp" exec 99>"$LOCK_FILE" flock -n 99 if [ "$?" != 0 ]; then exit 0 fi reload_dnsmasq_pids() { local pidfile pid find $TMP_PATH/acl -type f -name 'dnsmasq.pid' 2>/dev/null | while read pidfile; do if [ -s "$pidfile" ]; then read pid < "$pidfile" if [ -n "$pid" ] && kill -0 "$pid" 2>/dev/null; then kill -HUP "$pid" fi fi done } while true; do if [ -f "$LEASE_FILE" ]; then awk 'NF >= 4 && $4 != "*" {print $3" "$4}' "$LEASE_FILE" | sort > "$TMP_FILE" if [ -s "$TMP_FILE" ]; then if [ ! -f "$HOSTS_FILE" ] || ! cmp -s "$TMP_FILE" "$HOSTS_FILE"; then mv "$TMP_FILE" "$HOSTS_FILE" reload_dnsmasq_pids else rm -f "$TMP_FILE" fi else if [ -s "$HOSTS_FILE" ]; then : > "$HOSTS_FILE" reload_dnsmasq_pids fi rm -f "$TMP_FILE" fi fi sleep 60 done 2>/dev/null
294coder/Efficient-MIF
14,989
utils/_metric_legacy.py
##### # copy from https://github.com/XiaoXiao-Woo/PanCollection/blob/dev/UDL/pansharpening/common/evaluate.py # thanks a lot ##### import math import torch import torch.nn.functional as F import numpy as np from torch.autograd import Variable import decimal # 由于dat及其方差等数值舍入存在误差,最终结果有0.001左右的误差 def q2n(gt, x, q_blocks_size, q_shift): ''' ''' if isinstance(gt, torch.Tensor): gt = gt.cpu().numpy() x = x.cpu().numpy() N, N1, N2, N3 = gt.shape # 255 255 8 size2 = q_blocks_size # 32 stepx = math.ceil(N1 / q_shift) # 8 stepy = math.ceil(N2 / q_shift) # 8 if stepy <= 0: stepy = 1 stepx = 1 est1 = (stepx - 1) * q_shift + q_blocks_size - N1 # 1 est2 = (stepy - 1) * q_shift + q_blocks_size - N2 # 1 # if np.sum(np.array([est1 != 0, est2 != 0])) > 0: # refref = np.zeros(shape=[N1+1, N2+1]) # fusfus = refref.copy() for i in range(N3): a1 = gt[..., 0] ia1 = np.zeros(shape=[N, N1 + est1, N2 + est2]) ia1[:, : N1, : N2] = a1 ia1[:, :, N2:N2 + est2] = ia1[:, :, N2 - 1:-1:N2 - est2 + 1] ia1[:, N1:N1 + est1, ...] = ia1[:, N1 - 1:-1:N1 - est1 + 1, ...] if i == 0: refref = ia1[..., np.newaxis] # np.concatenate(refref, ia1, axis=3) else: refref = np.concatenate([refref, ia1[..., np.newaxis]], axis=-1) if i < N3: gt = gt[..., 1:] gt = refref for i in range(N3): a2 = x[..., 0] ia2 = np.zeros(shape=[N, N1 + est1, N2 + est2]) ia2[:, : N1, : N2] = a2 ia2[:, :, N2:N2 + est2] = ia2[:, :, N2 - 1:-1:N2 - est2 + 1] ia2[:, N1:N1 + est1, ...] = ia2[:, N1 - 1:-1:N1 - est1 + 1, ...] if i == 0: fusfus = ia2[..., np.newaxis] # np.concatenate(refref, ia1, axis=3) else: fusfus = np.concatenate([fusfus, ia2[..., np.newaxis]], axis=-1) if i < N3: x = x[..., 1:] x = fusfus x = np.array(x, dtype=np.uint16) gt = np.array(gt, dtype=np.uint16) _, N1, N2, N3 = gt.shape if math.ceil(math.log2(N3)) - math.log2(N3) != 0: Ndif = pow(2, math.ceil(math.log2(N3))) - N3 dif = np.zeros(shape=[N, N1, N2, Ndif], dtype=np.uint16) gt = np.concatenate(gt, dif, axis=-1) x = np.concatenate(x, dif, axis=-1) _, _, _, N3 = gt.shape valori = np.zeros(shape=[N, stepx, stepy, N3]) for j in range(stepx): for i in range(stepy): o = onions_quality(gt[:, j * q_shift:j * q_shift + q_blocks_size, i * q_shift: i * q_shift + size2, :], x[:, j * q_shift:j * q_shift + q_blocks_size, i * q_shift: i * q_shift + size2, :], q_blocks_size) # 0.971379489438014 0.00553590637316723 0.00305237797490489 -0.0188289323262161 -0.00420556598390016 -0.0173947468044076 -0.0202144450367593 0.0102693855205061 valori[:, j, i, :] = o q2n_idx_map = np.sqrt(np.sum(valori ** 2, axis=-1)) # q2n_index = np.mean(q2n_idx_map) return q2n_idx_map def norm_blocco(x, eps=1e-8): a = x.mean() c = x.std() if c == 0: c = eps return (x - a) / c + 1, a, c def onions_quality(dat1, dat2, size1): dat1 = np.float64(dat1) dat2 = np.float64(dat2) dat2 = np.concatenate([dat2[..., 0, np.newaxis], -dat2[..., 1:]], axis=-1) N, _, _, N3 = dat1.shape size2 = size1 # Block norm ''' 319.6474609375 37.05174450544686 357.970703125 61.54042371537683 518.708984375 111.53732768071865 608.23828125 154.26606056030568 593.412109375 163.97722215177643 554.8486328125 113.96758695803403 690.16015625 151.29524031046248 442.2314453125 94.12877724960003 mat 319.6475 37.0698 357.9707 61.5705 518.7090 111.5918 608.2383 154.3414 593.4121 164.0573 554.8486 114.0233 690.1602 151.3692 442.2314 94.1748 ''' for i in range(N3): a1, s, t = norm_blocco(np.squeeze(dat1[..., i])) # print(s,t) dat1[..., i] = a1 if s == 0: if i == 0: dat2[..., i] = dat2[..., i] - s + 1 else: dat2[..., i] = -(-dat2[..., i] - s + 1) else: if i == 0: dat2[..., i] = ((dat2[..., i] - s) / t) + 1 else: dat2[..., i] = -(((-dat2[..., i] - s) / t) + 1) m1 = np.zeros(shape=[N, N3]) m2 = m1.copy() mod_q1m = 0 mod_q2m = 0 mod_q1 = np.zeros(shape=[size1, size2]) mod_q2 = np.zeros(shape=[size1, size2]) for i in range(N3): m1[..., i] = np.mean(np.squeeze(dat1[..., i])) m2[..., i] = np.mean(np.squeeze(dat2[..., i])) mod_q1m += m1[..., i] ** 2 mod_q2m += m2[..., i] ** 2 mod_q1 += np.squeeze(dat1[..., i]) ** 2 mod_q2 += np.squeeze(dat2[..., i]) ** 2 mod_q1m = np.sqrt(mod_q1m) mod_q2m = np.sqrt(mod_q2m) mod_q1 = np.sqrt(mod_q1) mod_q2 = np.sqrt(mod_q2) termine2 = mod_q1m * mod_q2m # 7.97 termine4 = mod_q1m ** 2 + mod_q2m ** 2 # int1 = (size1 * size2) / (size1 * size2 - 1) * np.mean(mod_q1 ** 2) int2 = (size1 * size2) / (size1 * size2 - 1) * np.mean(mod_q2 ** 2) termine3 = int1 + int2 - (size1 * size2) / ((size1 * size2 - 1)) * (mod_q1m ** 2 + mod_q2m ** 2) # 17.8988 ** 2 mean_bias = 2 * termine2 / termine4 # 1 if termine3 == 0: q = np.zeros(shape=[N, 1, N3]) q[:, :, N3 - 1] = mean_bias else: cbm = 2 / termine3 # 32 32 8 qu = onion_mult2D(dat1, dat2) qm = onion_mult(m1.reshape(-1), m2.reshape(-1)) qv = np.zeros(shape=[N, N3]) for i in range(N3): qv[..., i] = (size1 * size2) / ((size1 * size2) - 1) * np.mean(np.squeeze(qu[:, :, i])) q = qv - (size1 * size2) / ((size1 * size2) - 1) * qm q = q * mean_bias * cbm return q def onion_mult2D(onion1, onion2): _, _, _, N3 = onion1.shape if N3 > 1: L = N3 // 2 a = onion1[..., : L] b = onion1[..., L:] b = np.concatenate([b[..., 0, np.newaxis], -b[..., 1:]], axis=-1) c = onion2[..., : L] d = onion2[..., L:] d = np.concatenate([d[..., 0, np.newaxis], -d[..., 1:]], axis=-1) if N3 == 2: ris = np.concatenate([a * c - d * b, a * d + c * b], axis=-1) else: ris1 = onion_mult2D(a, c) ris2 = onion_mult2D(d, np.concatenate([b[..., 0, np.newaxis], -b[..., 1:]], axis=-1)) ris3 = onion_mult2D(np.concatenate([a[..., 0, np.newaxis], -a[..., 1:]], axis=-1), d) ris4 = onion_mult2D(c, b) aux1 = ris1 - ris2 aux2 = ris3 + ris4 ris = np.concatenate([aux1, aux2], axis=-1) else: ris = onion1 * onion2 return ris def onion_mult(onion1, onion2): # _, N = onion1.shape N = len(onion1) if N > 1: L = N // 2 a = onion1[:L] b = onion1[L:] # b[1:] = -b[1:] b = np.append(np.array(b[0]), -b[1:]) c = onion2[:L] d = onion2[L:] # d[1:] = -d[1:] d = np.append(np.array(d[0]), -d[1:]) if N == 2: ris = np.append(a * c - d * b, a * d + c * b) else: ris1 = onion_mult(a, c) # b[1:] = -b[1:] ris2 = onion_mult(d, np.append(np.array(b[0]), -b[1:])) # a[1:] = -a[1:] ris3 = onion_mult(np.append(np.array(a[0]), -a[1:]), d) ris4 = onion_mult(c, b) aux1 = ris1 - ris2 aux2 = ris3 + ris4 ris = np.append(aux1, aux2) else: ris = np.array(onion1).reshape(-1) * np.array(onion2).reshape(-1) return ris def compute_index(img_base, img_out, ratio): h = img_out.shape[0] w = img_out.shape[1] chanel = img_out.shape[2] # 计算SAM sum1 = torch.sum(img_base * img_out, 2) sum2 = torch.sum(img_base * img_base, 2) sum3 = torch.sum(img_out * img_out, 2) t = (sum2 * sum3) ** 0.5 numlocal = torch.gt(t, 0) num = torch.sum(numlocal) t = sum1 / t angle = torch.acos(t) sumangle = torch.where(torch.isnan(angle), torch.full_like(angle, 0), angle).sum() if num == 0: averangle = sumangle else: averangle = sumangle / num SAM = averangle * 180 / 3.14159256 # 计算ERGAS summ = 0 for i in range(chanel): a1 = torch.mean((img_base[:, :, i] - img_out[:, :, i]) ** 2) m1 = torch.mean(img_base[:, :, i]) a2 = m1 * m1 summ = summ + a1 / a2 ERGAS = 100 * (1 / ratio) * ((summ / chanel) ** 0.5) return SAM, ERGAS decimal.getcontext().rounding = "ROUND_HALF_UP" n_digits = 6 # panHrnet: 2.6565 |1.4651 | 0.98364 | 0.98024 | 0.98089-Q8 def analysis_accu(img_base, img_out, ratio, flag_cut_bounds=True, dim_cut=1, choices=4): if flag_cut_bounds: img_base = img_base[dim_cut - 1:-dim_cut, dim_cut - 1:-dim_cut, :] #: img_out = img_out[dim_cut - 1:-dim_cut, dim_cut - 1:-dim_cut, :] #: # q2n # q2n_index = q2n(img_base, img_out, q_blocks_size=32, q_shift=32) h = img_out.shape[0] w = img_out.shape[1] chanel = img_out.shape[2] # 计算SAM sum1 = torch.sum(img_base * img_out, 2) sum2 = torch.sum(img_base * img_base, 2) sum3 = torch.sum(img_out * img_out, 2) t = (sum2 * sum3) ** 0.5 numlocal = torch.gt(t, 0) num = torch.sum(numlocal) t = sum1 / t angle = torch.acos(t) sumangle = torch.where(torch.isnan(angle), torch.full_like(angle, 0), angle).sum() if num == 0: averangle = sumangle else: averangle = sumangle / num # 或者采用https://segmentfault.com/a/1190000018929994修改精度 # averangle = math.ceil(averangle * 1000000) / 1000000 averangle = (averangle * 10 ** n_digits).round() / (10 ** n_digits) # SAM = decimal.Decimal(averangle.cpu().numpy() * 180 / 3.14159256).quantize(decimal.Decimal("0.00000")) SAM = averangle * 180 / 3.14159256 # 计算ERGAS summ = 0 for i in range(chanel): a1 = torch.mean((img_base[:, :, i] - img_out[:, :, i]) ** 2) m1 = torch.mean(img_base[:, :, i]) a2 = m1 * m1 summ = summ + a1 / a2 ERGAS = 100 * (1 / ratio) * ((summ / chanel) ** 0.5) # 计算PSNR mse = torch.mean((img_base - img_out) ** 2, 0) mse = torch.mean(mse, 0) rmse = mse ** 0.5 temp = torch.log(1 / rmse) / math.log(10) PSNR = 20 * temp # 计算SSIM # img_base = img_base.permute(2, 0, 1) # img_out = img_out.permute(2, 0, 1) # img_base = img_base.unsqueeze(0) # img_out = img_out.unsqueeze(0) # SSIM = _ssim(img_base.permute(2, 0, 1).unsqueeze(0), img_out.permute(2, 0, 1).unsqueeze(0)) # index = torch.zeros((5, chanel + 1)) # index[0, 1:chanel + 1] = CC # index[1, 1:chanel + 1] = PSNR # index[2, 1:chanel + 1] = SSIM # index[0, 0] = torch.mean(CC) # index[1, 0] = torch.mean(PSNR) # index[2, 0] = torch.mean(SSIM) # index[3, 0] = SAM # index[4, 0] = ERGAS PSNR = torch.mean(PSNR) # SSIM = torch.mean(SSIM) # q2n_index = np.mean(q2n_index) if choices == 5: # 计算CC C1 = torch.sum(torch.sum(img_base * img_out, 0), 0) - h * w * ( torch.mean(torch.mean(img_base, 0), 0) * torch.mean(torch.mean(img_out, 0), 0)) C2 = torch.sum(torch.sum(img_out ** 2, 0), 0) - h * w * (torch.mean(torch.mean(img_out, 0), 0) ** 2) C3 = torch.sum(torch.sum(img_base ** 2, 0), 0) - h * w * (torch.mean(torch.mean(img_base, 0), 0) ** 2) CC = C1 / ((C2 * C3) ** 0.5) CC = torch.mean(CC) return {'SAM': SAM, 'ERGAS': ERGAS, 'PSNR': PSNR, 'CC': CC} # , q2n_index return {'SAM': SAM, 'ERGAS': ERGAS, 'PSNR': PSNR, } def _ssim(img1, img2): img1 = img1.float() img2 = img2.float() channel = img1.shape[1] max_val = 1 _, c, w, h = img1.size() window_size = min(w, h, 11) sigma = 1.5 * window_size / 11 window = create_window(window_size, sigma, channel).cuda() mu1 = F.conv2d(img1, window, padding=window_size // 2, groups=channel) mu2 = F.conv2d(img2, window, padding=window_size // 2, groups=channel) mu1_sq = mu1.pow(2) mu2_sq = mu2.pow(2) mu1_mu2 = mu1 * mu2 sigma1_sq = F.conv2d(img1 * img1, window, padding=window_size // 2, groups=channel) - mu1_sq sigma2_sq = F.conv2d(img2 * img2, window, padding=window_size // 2, groups=channel) - mu2_sq sigma12 = F.conv2d(img1 * img2, window, padding=window_size // 2, groups=channel) - mu1_mu2 C1 = (0.01 * max_val) ** 2 C2 = (0.03 * max_val) ** 2 V1 = 2.0 * sigma12 + C2 V2 = sigma1_sq + sigma2_sq + C2 ssim_map = ((2 * mu1_mu2 + C1) * V1) / ((mu1_sq + mu2_sq + C1) * V2) t = ssim_map.shape return ssim_map.mean(2).mean(2) def gaussian(window_size, sigma): gauss = torch.Tensor([math.exp(-(x - window_size // 2) ** 2 / float(2 * sigma ** 2)) for x in range(window_size)]) return gauss / gauss.sum() def create_window(window_size, sigma, channel): _1D_window = gaussian(window_size, sigma).unsqueeze(1) _2D_window = _1D_window.mm(_1D_window.t()).float().unsqueeze(0).unsqueeze(0) window = Variable(_2D_window.expand(channel, 1, window_size, window_size).contiguous()) return window def compare_index(A): A_size = A.shape ite_n = A_size[2] band_n = A_size[1] C_better = A[:, 0, 0] ind = 0 for i in range(ite_n): score_b = 0 score_c = 0 C_compare = A[:, 0, i] if (C_better[0] > C_compare[0]): score_b = score_b + 1 else: score_c = score_c + 1 if (C_better[1] > C_compare[1]): score_b = score_b + 1 else: score_c = score_c + 1 if (C_better[2] > C_compare[2]): score_b = score_b + 1 else: score_c = score_c + 1 if (C_better[3] < C_compare[3]): score_b = score_b + 1 else: score_c = score_c + 1 if (C_better[4] < C_compare[4]): score_b = score_b + 1 else: score_c = score_c + 1 if (score_c > score_b): C_better = A[:, 0, i] ind = i C_best = A[:, :, ind] best_ind = ind + 1 return C_best, best_ind ######################################## # Non-reference quality assessment # D_\lambda, D_s, HQNR implementaion # Author: Xiao Wu, Zihan Cao # Date: 2024/02/04 ######################################## from .sharpening_index.hqnr2 import indexes_evaluation_FS if __name__ == "__main__": a = torch.randn(256, 256, 3) b = torch.randn(256, 256, 3) print(analysis_accu(a, b, 1, choices=5))
294coder/Efficient-MIF
9,158
utils/network_utils.py
import math from typing import Union from copy import deepcopy from bytecode import Bytecode, Instr from accelerate import Accelerator from accelerate.utils import DistributedType import torch import torch.nn as nn def hook_model(model: nn.Module, saved_tensor, hook_class): def feature_hook(_, input, output): # forward hook saved_tensor.append([input, output]) hooks = [] for n, m in model.named_modules(): if isinstance(m, hook_class): hooks.append(m.register_forward_hook(feature_hook)) return model, hooks class loss_with_l2_regularization(nn.Module): def __init__(self): super(loss_with_l2_regularization, self).__init__() def forward(self, loss, model, weight_decay=1e-5, flag=False): regularizations = [] for k, v in model.named_parameters(): if "conv" in k and "weight" in k: # print(k) penality = weight_decay * ((v.data ** 2).sum() / 2) regularizations.append(penality) if flag: print("{} : {}".format(k, penality)) # r = torch.sum(regularizations) loss = loss + sum(regularizations) return loss def variance_scaling_initializer(tensor): # stole it from woo-xiao. # thanks def calculate_fan(shape, factor=2.0, mode="FAN_IN", uniform=False): # 64 9 3 3 -> 3 3 9 64 # 64 64 3 3 -> 3 3 64 64 if shape: # fan_in = float(shape[1]) if len(shape) > 1 else float(shape[0]) # fan_out = float(shape[0]) fan_in = float(shape[-2]) if len(shape) > 1 else float(shape[-1]) fan_out = float(shape[-1]) else: fan_in = 1.0 fan_out = 1.0 for dim in shape[:-2]: fan_in *= float(dim) fan_out *= float(dim) if mode == "FAN_IN": # Count only number of input connections. n = fan_in elif mode == "FAN_OUT": # Count only number of output connections. n = fan_out elif mode == "FAN_AVG": # Average number of inputs and output connections. n = (fan_in + fan_out) / 2.0 if uniform: raise NotImplemented # # To get stddev = math.sqrt(factor / n) need to adjust for uniform. # limit = math.sqrt(3.0 * factor / n) # return random_ops.random_uniform(shape, -limit, limit, # dtype, seed=seed) else: # To get stddev = math.sqrt(factor / n) need to adjust for truncated. trunc_stddev = math.sqrt(1.3 * factor / n) return fan_in, fan_out, trunc_stddev def model_params(model: nn.Module, accelerator=None): if accelerator is not None: model = accelerator.unwrap_model(model) if accelerator.distributed_type == DistributedType.DEEPSPEED: from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint return get_fp32_state_dict_from_zero_checkpoint(model) else: return model.state_dict() if isinstance(model, (nn.DataParallel, nn.parallel.DistributedDataParallel)): model = model.module elif isinstance(model, torch._dynamo.eval_frame.OptimizedModule): # torch.compile model model = model._orig_mod return model.state_dict() def model_device(model: Union[nn.Module, nn.DataParallel, nn.parallel.DistributedDataParallel, torch._dynamo.eval_frame.OptimizedModule]): params = model.parameters() p0 = next(params) return p0.device def clip_norm_(max_norm, network, fp_scaler=None, optim=None, accelerator=None): if fp_scaler is not None: fp_scaler.unscale_(optim) if accelerator is not None: accelerator.clip_grad_norm_(network.parameters(), max_norm) else: torch.nn.utils.clip_grad.clip_grad_norm_(network.parameters(), max_norm) def clip_value_(max_value, network, fp_scaler=None, optim=None, accelerator=None): if fp_scaler is not None: fp_scaler.unscale_(optim) if accelerator is not None: accelerator.clip_grad_value(network.parameters(), max_value) else: torch.nn.utils.clip_grad.clip_grad_value_(network.parameters(), max_value) def step_loss_backward( optim, network=None, max_norm=None, max_value=None, loss=None, fp16=False, # decrepted mix_precision=False, fp_scaler=None, grad_accum=False, accelerator: Accelerator=None, ): """ :param optim: optimizer. type: optim.Optimizer :param network: instanced network. type: nn.Module :param max_norm: clip norm. type: float :param loss: float :param fp16: bool :param fp_scaler: mix-precision scaler :return: """ if (fp16 and fp_scaler is None) or (not fp16 and fp_scaler is not None): if accelerator is None: raise ValueError("fp16 and grad_scaler should be set together") else: fp16 = False if max_norm is not None and network is None: raise ValueError("max_norm is set, network should be set") mixed_precision = fp16 optim.zero_grad() if mixed_precision: fp_scaler.scale(loss).backward() if max_norm is not None: clip_norm_(max_norm, network, fp_scaler, optim) if not grad_accum: fp_scaler.step(optim) fp_scaler.update() else: if accelerator is None: loss.backward() else: accelerator.backward(loss) # assert max_norm and max_value can not be set at the same time if max_norm is not None: clip_norm_(max_norm, network, accelerator=accelerator) elif max_value is not None: clip_value_(network.parameters(), max_value, accelerator=accelerator) if not grad_accum: optim.step() def find_no_or_big_norm_params(net: nn.Module, ktop: int=20): """find unused params and big-normed gradient Args: net (nn.Module): network to be checked. ktop (int, optional): top k params to be printed. Defaults to 20. """ # find unused params and big-normed gradient d_grads = {} n_params = 0 for n, p in net.named_parameters(): n_params += p.numel() if p.grad is None: print(n, "has no grad") else: p_sum = torch.abs(p.grad).sum().item() d_grads[n] = p_sum # topk d_grads = dict(sorted(d_grads.items(), key=lambda item: item[1], reverse=True)) for k, v in list(d_grads.items())[:ktop]: print(k, v) class EMAModel(object): def __init__(self, model, ema_ratio=0.9999): super().__init__() self.model = model self.ema_ratio = ema_ratio self.ema_model = deepcopy(model) def update(self): for ema_p, now_p in zip(self.ema_model.state_dict(), self.model.state_dict()): ema_p.data = ema_p.data * self.ema_ratio + now_p.data * (1 - self.ema_ratio) def ema_model_state_dict(self): try: return self.ema_model.module.state_dict() except: return self.ema_model.state_dict() class get_local(object): cache = {} is_activate = False def __init__(self, *args): self.varname = args def __call__(self, func): if not type(self).is_activate: return func type(self).cache[func.__qualname__] = [] c = Bytecode.from_code(func.__code__) extra_code = [] extra_code.extend([ *[Instr('LOAD_FAST', varn) for varn in self.varname], Instr('BUILD_LIST', len(self.varname)), Instr('STORE_FAST', '_result_list'), Instr('LOAD_FAST', '_result_list'), Instr('BUILD_TUPLE', 2) ]) c[-1:-1] = extra_code func.__code__ = c.to_code() def wrapper(*args, **kwargs): res = func(*args, **kwargs) output_vs = res[0] saved_vs = res[1] # assume is a Tensor for i, v in enumerate(saved_vs): if hasattr(v, 'detach'): v = v.detach().cpu() #.numpy() saved_vs[i] = v type(self).cache[func.__qualname__].append(saved_vs) return output_vs return wrapper @classmethod def clear(cls): for key in cls.cache.keys(): cls.cache[key] = [] @classmethod def activate(cls): from loguru import logger logger.warning('ready to get local varibles, be careful about you CPU memory.') cls.is_activate = True # if __name__ == '__main__': # get_local.activate() # @get_local('x', 'y') # def func(): # x = 1 # y = [1,2] # return x, y # print(func()) # print(get_local.cache)
294coder/Efficient-MIF
10,596
utils/metric.py
import numpy as np import torch from skimage.metrics import structural_similarity, peak_signal_noise_ratio from functools import partial from warnings import warn import sys sys.path.append('./') from utils.misc import to_numpy from utils.misc import dict_to_str, to_numpy from utils._metric_legacy import analysis_accu, indexes_evaluation_FS class NonAnalysis(object): def __init__(self): self.acc_ave = {} # only used as attribution pass def __call__(self, *args, **kwargs): pass def __repr__(self): return 'NonAnalysis()' # TODO: need to be tested the new metric analysis # FIXME: this python code is not same as matlab code, you should use matlab code to get the real accuracy # only used in training and validate class AnalysisPanAcc(object): def __init__(self, ratio=4, ref=True, ergas_ratio: int = 4, **unref_factory_kwargs): """pansharpening metric analysis class Args: ratio (int, optional): fusion ratio. Defaults to 4. ref (bool, optional): reduce-resolution or full-resolution. Defaults to True. ergas_ratio (int, optional): previous api (may decrepated soon). Defaults to 4. unref_factory_kwargs(dict): sensor, default_max_value. Defaults to {'sensor': 'default', 'default_max_value': None}. Raises: ValueError: _description_ """ # ergas_ratio is decrepated if ratio is None: ratio = ergas_ratio warn('@ergas_ratio is deprecated, use ratio instead') self.ratio = ratio self.ref = ref # metric functions if ref: self.__sam_ergas_psnr_cc_one_image = partial(analysis_accu, ratio=ergas_ratio, choices=5) self.ssim = ssim_batch_tensor_metric else: # @sensor in ['QB', 'IKONOS', 'WV2', 'WV3', 'default'] assert 'sensor' in unref_factory_kwargs or 'default_max_value' in unref_factory_kwargs, \ '@sensor or @default_max_value should be specified in unrefactory_kwargs' sensor = unref_factory_kwargs.pop('sensor', 'default').upper() if sensor == 'DEFAULT': warn('sensor is not specified, use default sensor type') self.default_max_value = unref_factory_kwargs.pop('default_max_value', None) if self.default_max_value is None: _default_max_value = {'QB': 2047, 'IKONOS': 1023, 'WV2': 2047, 'WV3': 2047, 'GF2': 1023, 'DEFAULT': 2047, 'CAVE_X4': 1, 'CAVE_X8': 1, 'HARVARD_X': 1, 'HARVARD_X8': 1, 'GF5':1, 'GF2-GF5': 1,} self.default_max_value = _default_max_value.get(sensor) print(f'>>> @default_max_value is not specified, set it according to @sensor:' f'{sensor, self.default_max_value}\n', '-'*20) self.FS_metric_fn = partial(indexes_evaluation_FS, L=11, Qblocks_size=32, sensor=sensor, th_values=0, ratio=ratio, flagQNR=False) # tracking accuracy self._acc_d = {} self._call_n = 0 self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if ref else \ {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} @property def empty_acc(self): return {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \ {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} @staticmethod def permute_dim(*args, permute_dims=(1, 2, 0)): l = [] for i in args: l.append(i.permute(*permute_dims)) return l @staticmethod def _sum_acc(d_ave, d_now, n, n2=1): assert len(d_ave) == len(d_now) for k in d_ave.keys(): v2 = d_now[k] * n2 d_ave[k] *= n d_ave[k] += v2.cpu().item() if isinstance(v2, torch.Tensor) else v2 return d_ave @staticmethod def _average_acc(d_ave, n): for k in d_ave.keys(): d_ave[k] /= n return d_ave def sam_ergas_psnr_cc_batch(self, b_gt, b_pred): n = b_gt.shape[0] # input shape should be [B, C, H, W] acc_ds = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0.} for i, (img1, img2) in enumerate(zip(b_gt, b_pred)): img1, img2 = self.permute_dim(img1, img2) acc_d = self.__sam_ergas_psnr_cc_one_image(img1, img2) acc_ds = self._sum_acc(acc_ds, acc_d, i) acc_ds = self._average_acc(acc_ds, i + 1) return acc_ds def D_lambda_D_s_HQNR_batch(self, sr=None, ms=None, lms=None, pan=None): assert sr is not None and lms is not None and pan is not None and ms is not None if ms is None: ms = torch.nn.functional.interpolate(lms, scale_factor=1/self.rato, mode='bilinear', align_corners=False) acc_ds = {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} sr, ms, lms, pan = self.permute_dim(sr, ms, lms, pan, permute_dims=(0, 2, 3, 1)) sr, ms, lms, pan = to_numpy(sr, ms, lms, pan) _max_value = getattr(self, 'default_max_value') sr, ms, lms, pan = map(lambda x: np.clip(x * _max_value, 0, _max_value), [sr, ms, lms, pan]) for i, (sr_i, ms_i, lms_i, pan_i) in enumerate(zip(sr, ms, lms, pan)): QNR_index, D_lambda, D_S, _ = self.FS_metric_fn(I_F=sr_i, I_MS_LR=ms_i, I_MS=lms_i, I_PAN=pan_i) acc_d = dict(HQNR=QNR_index, D_lambda=D_lambda, D_S=D_S) acc_ds = self._sum_acc(acc_ds, acc_d, i) acc_ds = self._average_acc(acc_ds, i + 1) return acc_ds def once_batch_call(self, **kwargs): if self.ref: acc_d1 = self.sam_ergas_psnr_cc_batch(**kwargs) acc_ssim = self.ssim(**kwargs) acc_d1['SSIM'] = acc_ssim else: acc_d1 = self.D_lambda_D_s_HQNR_batch(**kwargs) self._acc_d = acc_d1 return acc_d1 def _call_check_args_to_kwargs(self, *args): def may_np_to_tensor(d): for k, v in d.items(): if not isinstance(v, torch.Tensor): d[k] = torch.tensor(v, dtype=torch.float32) elif v.dtype != torch.float32: d[k] = v.float() return d if len(args) == 2: assert self.ref, 'ref mode should have 2 args' kwargs = dict(b_gt=args[0], b_pred=args[1]) elif len(args) == 3: assert not self.ref, 'unref mode should have more than 2 args' kwargs = dict(sr=args[0], lms=args[1], pan=args[2]) elif len(args) == 4: assert not self.ref, 'unref mode should have more than 2 args' kwargs = dict(sr=args[0], ms=args[1], lms=args[2], pan=args[3]) else: raise ValueError('args should have 2 or 4 elements') return may_np_to_tensor(kwargs) def __call__(self, *args): """ Args: ref mode: b_gt (torch.Tensor): [b, c, h, w] b_pred (torch.Tensor): [b, c, h, w] unref mode: sr (torch.Tensor): [b, c, h, w] ms (torch.Tensor, optional): [b, c, h/ratio, w/ratio] lms (torch.Tensor): [b, c, h, w] pan (torch.Tensor): [b, c, h, w] """ kwargs = self._call_check_args_to_kwargs(*args) n = args[0].shape[0] self.acc_ave = self._sum_acc(self.acc_ave, self.once_batch_call(**kwargs), self._call_n, n2=n) self.acc_ave = self._average_acc(self.acc_ave, self._call_n + n) self._call_n += n return self.acc_ave def clear_history(self, verbose=False): if verbose: print('>> AccAnalysis: clear history') self._acc_d = {} self._call_n = 0 self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \ {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.} def print_str(self, decimals=6): return dict_to_str(self.acc_ave, decimals=decimals) def __repr__(self) -> str: repr_str = f'AnalysisPanAcc(ratio={self.ratio}, ref={self.ref}):' repr_str += f'\n{self.print_str()}' return repr_str def normalize_to_01(x): # normalize tensor to [0, 1] if isinstance(x, torch.Tensor): x -= x.flatten(-2).min(-1, keepdim=True)[0][..., None] x /= x.flatten(-2).max(-1, keepdim=True)[0][..., None] elif isinstance(x, np.ndarray): x -= x.min((-2, -1), keepdims=True) x /= x.max((-2, -1), keepdims=True) return x def psnr_one_img(img_gt, img_test): """ calculate PSNR for one image :param img_gt: ground truth image, numpy array, shape [H, W, C] :param img_test: test or inference image, numpy array, shape [H, W, C] :return: PSNR, float type """ assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size' return peak_signal_noise_ratio(img_gt, img_test) def psnr_batch_tensor_metric(b_gt, b_pred): """ calculate PSNR for batch tensor images :param b_gt: tensor, shape [B, C, H, W] :param b_test: tensor, shape [B, C, H, W] :return: """ assert b_gt.shape[0] == b_pred.shape[0] bs = b_gt.shape[0] psnr = 0. for gt, t in zip(b_gt, b_pred): psnr += psnr_one_img(*(to_numpy(gt, t))) return psnr / bs def ssim_one_image(img_gt, img_test, channel_axis=0): assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size' return structural_similarity(img_gt, img_test, channel_axis=channel_axis, data_range=1.) def ssim_batch_tensor_metric(b_gt, b_pred): assert b_gt.shape[0] == b_pred.shape[0] bs = b_gt.shape[0] ssim = 0. for gt, t in zip(b_gt, b_pred): ssim += ssim_one_image(*(to_numpy(gt, t)), channel_axis=0) return ssim / bs if __name__ == '__main__': sr = torch.rand(4, 3, 256, 256) ms = torch.rand(4, 3, 64, 64) lms = torch.rand(4, 3, 256, 256) pan = torch.rand(4, 3, 256, 256) gt = torch.rand(4, 3, 256, 256) analysis = AnalysisPanAcc(ref=False, ratio=4, default_max_value=2047) for i in range(2): analysis(sr[i:i+2], ms[i:i+2], lms[i:i+2], pan[i:i+2]) print(analysis.print_str())
2977094657/DidaAPI
1,671
frontend/docs/api/auth/index.md
# 认证概览 滴答清单API提供两种认证方式,您可以根据使用场景选择合适的登录方法。 ## 认证方式对比 | 特性 | 密码登录 | 微信登录 | |------|----------|----------| | **便利性** | 需要输入用户名密码 | 扫码即可 | | **安全性** | 依赖密码强度 | 依赖微信安全体系 | | **适用场景** | 自动化脚本、批量操作 | 交互式使用 | | **实现复杂度** | 简单直接 | 复杂(需要轮询) | | **网络要求** | 一次HTTP请求 | 多次请求和轮询 | | **用户体验** | 传统表单登录 | 现代扫码体验 | ## 认证流程 ### 密码登录流程 ```mermaid sequenceDiagram participant User as 用户 participant Client as 客户端 participant API as API服务器 participant Dida as 滴答清单服务器 User->>Client: 1. 输入用户名密码 Client->>API: 2. POST https://api.dida365.com/api/v2/user/signon API->>Dida: 3. 验证登录凭据 Dida-->>API: 4. 返回认证令牌 API->>API: 5. 保存会话信息 API-->>Client: 6. 返回登录结果 Client->>User: 7. 显示登录状态 ``` ### 微信登录流程 ```mermaid sequenceDiagram participant User as 用户 participant Client as 客户端 participant API as API服务器 participant WeChat as 微信服务器 participant Dida as 滴答清单服务器 User->>Client: 1. 请求微信登录 Client->>API: 2. GET https://open.weixin.qq.com/connect/qrconnect API->>WeChat: 3. 获取二维码 WeChat-->>API: 4. 返回二维码 API-->>Client: 5. 返回二维码URL Client->>User: 6. 显示二维码 User->>WeChat: 7. 微信扫码确认 Client->>API: 8. GET https://lp.open.weixin.qq.com/connect/l/qrconnect API->>WeChat: 9. 轮询登录状态 WeChat-->>API: 10. 返回授权码 API->>Dida: 11. 验证授权码 Dida-->>API: 12. 返回认证令牌 API->>API: 13. 保存会话信息 API-->>Client: 14. 返回登录结果 ``` ## 详细文档 ### 密码登录相关 - [密码登录](./password-login.md) - 详细的密码登录接口文档 ### 微信登录相关 - [微信登录流程](./wechat-login-flow.md) - 完整的微信扫码登录流程 - [获取微信二维码](./get-wechat-qrcode.md) - 获取微信登录二维码 - [轮询登录状态](./poll-login-status.md) - 检查二维码扫码状态 - [验证微信登录](./validate-wechat-login.md) - 验证微信登录并获取令牌
294coder/Efficient-MIF
20,170
utils/misc.py
import argparse from contextlib import contextmanager import json from multiprocessing import context import os import os.path as osp import random import time from typing import Dict, Iterable, Sequence, Union import importlib import h5py import kornia.augmentation as K from fvcore.nn import FlopCountAnalysis, flop_count_table import yaml import numpy as np import torch import torch.distributed as dist import kornia import shortuuid from matplotlib import pyplot as plt from torch.backends import cudnn def exists(val): return val is not None def default(val, d): return val if exists(val) else d def is_none(val): return val in ('none', 'None', 'NONE', None) def set_all_seed(seed=2022): torch.manual_seed(seed) torch.cuda.manual_seed(seed) torch.cuda.manual_seed_all(seed) random.seed(seed) np.random.seed(seed) cudnn.deterministic = True cudnn.benchmark = False def to_numpy(*args): l = [] for i in args: if isinstance(i, torch.Tensor): l.append(i.detach().cpu().numpy()) return l def to_tensor(*args, device, dtype): out = [] for a in args: out.append(torch.tensor(a, dtype=dtype).to(device)) return out def args_no_str_none(value: str) -> "str | None": if value.lower() == "none": return None return value def to_device(*args, device): out = [] for a in args: out.append(a.to(device)) return out def rgb_to_ycbcr(image: torch.Tensor) -> torch.Tensor: """ Convert an RGB image to YCbCr. Args: image: RGB image tensor with shape (..., 3, H, W) in range [0, 1] Returns: YCbCr image tensor with shape (..., 3, H, W) """ if not isinstance(image, torch.Tensor): raise TypeError(f"Input type is not a torch.Tensor. Got {type(image)}") if len(image.shape) < 3 or image.shape[-3] != 3: raise ValueError(f"Input size must have a shape of (..., 3, H, W). Got {image.shape}") r: torch.Tensor = image[..., 0, :, :] g: torch.Tensor = image[..., 1, :, :] b: torch.Tensor = image[..., 2, :, :] y: torch.Tensor = 0.29900 * r + 0.58700 * g + 0.11400 * b cb: torch.Tensor = -0.168736 * r - 0.331264 * g + 0.50000 * b + 0.5 cr: torch.Tensor = 0.50000 * r - 0.418688 * g - 0.081312 * b + 0.5 return torch.stack([y, cb, cr], dim=-3) def ycbcr_to_rgb(image: torch.Tensor) -> torch.Tensor: """ Convert a YCbCr image to RGB. Args: image: YCbCr image tensor with shape (..., 3, H, W) Returns: RGB image tensor with shape (..., 3, H, W) in range [0, 1] """ if not isinstance(image, torch.Tensor): raise TypeError(f"Input type is not a torch.Tensor. Got {type(image)}") if len(image.shape) < 3 or image.shape[-3] != 3: raise ValueError(f"Input size must have a shape of (..., 3, H, W). Got {image.shape}") y: torch.Tensor = image[..., 0, :, :] cb: torch.Tensor = image[..., 1, :, :] cr: torch.Tensor = image[..., 2, :, :] r: torch.Tensor = y + 1.40200 * (cr - 0.5) g: torch.Tensor = y - 0.34414 * (cb - 0.5) - 0.71414 * (cr - 0.5) b: torch.Tensor = y + 1.77200 * (cb - 0.5) return torch.stack([r, g, b], dim=-3).clamp(0, 1) @contextmanager def y_pred_model_colored(vis: torch.Tensor, enable: bool=True): """ Context manager to handle YCbCr color space conversion for image processing. Args: vis (torch.Tensor): Input RGB image tensor of shape (B, 3, H, W) Yields: torch.Tensor: Y channel of the image The context manager handles: 1. Converting RGB to YCbCr 2. Extracting Y channel for processing 3. Converting processed Y channel back to RGB """ if enable: assert vis.size(1) == 3, 'vis should be a 3-channel rgb image' y_cb_cr = kornia.color.rgb_to_ycbcr(vis) cbcr = y_cb_cr[:, 1:] y = y_cb_cr[:, :1] def back_to_rgb(pred_y): y_cb_cr = torch.cat([pred_y, cbcr], dim=1) return kornia.color.ycbcr_to_rgb(y_cb_cr) else: y = vis def back_to_rgb(pred_rgb): return pred_rgb try: # Yield the Y channel for processing yield y, back_to_rgb finally: pass class WindowBasedPadder(object): def __init__(self, window_size=64) -> None: self.window_size = window_size self.padding_fn = None def find_least_pad(self, base_size: tuple, window_size: int): least_size = [] for b_s in base_size: if b_s % window_size == 0: least_size.append(b_s) else: mult = b_s // window_size mult += 1 least_size.append(mult * window_size) return least_size def __call__(self, img: torch.Tensor, size: Sequence[int]=None, no_check_pad: bool = False): if no_check_pad: assert self.padding_fn is not None return self.padding_fn(img) if size is not None: self._last_img_ori_size = size self.padding_fn = K.PadTo(size) else: pad_size = self.find_least_pad(img.shape[-2:], self.window_size) self._last_img_ori_size = img.shape[-2:] self.padding_fn = K.PadTo(pad_size) return self.padding_fn(img) def inverse(self, img: torch.Tensor): return self.padding_fn.inverse(img, size=self._last_img_ori_size) def h5py_to_dict(file: h5py.File, keys=None) -> dict[str, np.ndarray]: """get all content in a h5py file into a dict contains key and values Args: file (h5py.File): h5py file keys (list, optional): h5py file keys used to extract values. Defaults to ["ms", "lms", "pan", "gt"]. Returns: dict[str, np.ndarray]: """ d = {} if keys is None: keys = list(file.keys()) for k in keys: v = file[k][:] d[k] = v return d def dict_to_str(d, decimals=4): n = len(d) func = lambda k, v: f"{k}: {torch.round(v, decimals=decimals).item() if isinstance(v, torch.Tensor) else np.round(v, decimals=decimals)}" s = "" for i, (k, v) in enumerate(d.items()): s += func(k, v) + (", " if i < n - 1 else "") return s def prefixed_dict_key(d, prefix, sep="_"): # e.g. # SSIM -> train_SSIM d2 = {} for k, v in d.items(): d2[prefix + sep + k] = v return d2 # TODO: nees test class CheckPointManager(object): def __init__( self, model: torch.nn.Module, save_path: str, save_every_eval: bool = False, verbose: bool = True, ): """ manage model checkpoints Args: model: nn.Module, can be single node model or multi-nodes model save_path: str like '/home/model_ckpt/resnet.pth' or '/home/model_ckpt/exp1' when @save_every_eval is False or True save_every_eval: when False, save params only when ep_loss is less than optim_loss. when True, save params every eval epoch verbose: print out all information e.g. @save_every_eval=False, @save_path='/home/ckpt/resnet.pth' weights will be saved like ------------- /home/ckpt |-resnet.pth ------------- @save_every_eval=True, @save_path='/home/ckpt/resnet' weights will be saved like ------------- /home/ckpt |-resnet |-ep_20.pth |-ep_40.pth ------------- """ self.model = model self.save_path = save_path self.save_every_eval = save_every_eval self._optim_loss = torch.inf self.verbose = verbose self.check_path_legal() def check_path_legal(self): if self.save_every_eval: if not os.path.exists(self.save_path): os.makedirs(self.save_path) else: assert self.save_path.endswith(".pth") par_dir = os.path.dirname(self.save_path) if not os.path.exists(par_dir): os.makedirs(par_dir) def save( self, ep_loss: Union[float, torch.Tensor] = None, ep: int = None, extra_saved_dict: dict = None, ): """ Args: ep_loss: should be set when @save_every_eval=False ep: should be set when @save_every_eval=True extra_saved_dict: a dict which contains other information you want to save with model e.g. {'optimizer_ckpt': op_ckpt, 'time': '2023/1/21'} Returns: """ if isinstance(ep_loss, torch.Tensor): ep_loss = ep_loss.item() saved_dict = {} if not self.save_every_eval: assert ep_loss is not None if ep_loss < self._optim_loss: self._optim_loss = ep_loss path = self.save_path saved_dict["optim_loss"] = ep_loss else: print( "optim loss: {}, now loss: {}, not saved".format( self._optim_loss, ep_loss ) ) return else: assert ep is not None path = os.path.join(self.save_path, "ep_{}.pth".format(ep)) if extra_saved_dict is not None: assert "model" not in list(saved_dict.keys()) saved_dict = extra_saved_dict try: saved_dict["model"] = self.model.module.state_dict() except: saved_dict["model"] = self.model.state_dict() torch.save(saved_dict, path) if self.verbose: print( f"saved params contains\n", *[ "\t -{}: {}\n".format(k, v if k != "model" else "model params") for k, v in saved_dict.items() ], "saved path: {}".format(path), ) def is_main_process(func=None): """ check if current process is main process in ddp warning: if not in ddp mode, always return True :return: """ def _is_main_proc(): if dist.is_initialized(): return dist.get_rank() == 0 else: return True if func is None: return _is_main_proc() else: def warp_func(*args, **kwargs): if _is_main_proc(): return func(*args, **kwargs) else: return None return warp_func def print_args(args): d = args.__dict__ for k, v in d.items(): print(f"{k}: {v}") def yaml_load(name, base_path="./configs", end_with="_config.yaml"): path = osp.join(base_path, name + end_with) if osp.exists(path): f = open(path) cont = f.read() return yaml.load(cont, Loader=yaml.FullLoader) else: print("configuration file not exists") raise FileNotFoundError(f'file not exists: {path}') def json_load(name, base_path="./configs"): path = osp.join(base_path, name + "_config.json") with open(path) as f: return json.load(f) def config_py_load(name, base_path="configs"): args = importlib.import_module(f".{name}_config", package=base_path) return args.config class NameSpace: def __init__(self, **kwargs): self.__dict__.update(kwargs) @property def attrs(self): return self.__dict__ def to_dict(self): out = {} d = self.attrs for k, v in d.items(): if isinstance(v, NameSpace): out[k] = v.to_dict() else: out[k] = v return out def __repr__(self, d=None, nprefix=0): repr_str = "" if d is None: d = self.attrs for k, v in d.items(): if isinstance(v, NameSpace): repr_str += ( " " * nprefix + f"{k}: \n" + f"{self.__repr__(v.attrs, nprefix + 1)}" ) else: repr_str += " " * nprefix + f"{k}: {v}\n" return repr_str def __getitem__(self, item): return self.attrs[item] def __setitem__(self, key, value): setattr(self.attrs, key, value) def recursive_search_dict2namespace(d: Dict): """ convert a yaml-like configuration (dict) to namespace-like class e.g. {'lr': 1e-3, 'path': './datasets/train_wv3.h5'} -> NameSpace().lr = 1e-3, NameSpace().path = './datasets/train_wv3.h5' Warning: the value in yaml-like configuration should not be another dict :param d: :return: """ namespace = NameSpace() for k, v in d.items(): if isinstance(v, dict): setattr(namespace, k, recursive_search_dict2namespace(v)) else: setattr(namespace, k, v) return namespace def merge_args_namespace(parser_args: argparse.Namespace, namespace_args: NameSpace): """ merge parser_args and self-made class _NameSpace configurations together for better usage. return args that support dot its member, like args.optimizer.lr :param parser_args: :param namespace_args: :return: """ # namespace_args.__dict__.update(parser_args.__dict__) namespace_d = namespace_args.__dict__ for k, v in parser_args.__dict__.items(): if not (k in namespace_d.keys() and v is None): setattr(namespace_args, k, v) return namespace_args def generate_id(length: int = 8) -> str: # ~3t run ids (36**8) run_gen = shortuuid.ShortUUID(alphabet=list("0123456789abcdefghijklmnopqrstuvwxyz")) return str(run_gen.random(length)) def find_weight(weight_dir="./weight/", id=None, func=None): """ return weight absolute path referring to id Args: weight_dir: weight dir that saved weights id: weight id func: split string function Returns: str, absolute path """ assert id is not None, "@id can not be None" weight_list = os.listdir(weight_dir) if func is None: func = lambda x: x.split(".")[0].split("_")[-1] for id_s in weight_list: only_id = func(id_s) if only_id == id: return os.path.abspath(os.path.join(weight_dir, id_s)) print(f"can not find {id}") return None def _delete_unneeded_weight_file(weight_dir="./weight/", id=None): """ delete unneeded weight file referring to id Args: weight_dir: id: Returns: """ assert id is not None, "@id can not be None" abspath = find_weight(weight_dir, id) if abspath is not None: assert os.path.exists(abspath) os.remove(abspath) print(f"delete {os.path.basename(abspath)}") def print_network_params_macs_fvcore(network, *inputs): """ print out network's parameters and macs by using fvcore package Args: network: nn.Module *inputs: input argument Returns: None """ analysis = FlopCountAnalysis(network, inputs=inputs) print(flop_count_table(analysis)) def clip_dataset_into_small_patches( file: h5py.File, patch_size: int, up_ratio: int, ms_channel: int, pan_channel: int, dataset_keys: Union[list[str], tuple[str]] = ("gt", "ms", "lms", "pan"), save_path: str = "./data/clip_data.h5", ): """ clip patches at spatial dim Args: file: h5py.File of original dataset patch_size: ms clipped size up_ratio: shape of lms divide shape of ms ms_channel: pan_channel: dataset_keys: similar to [gt, ms, lms, pan] save_path: must end with h5 Returns: """ unfold_fn = lambda x, c, ratio: ( torch.nn.functional.unfold( x, kernel_size=patch_size * ratio, stride=patch_size * ratio ) .transpose(1, 2) .reshape(-1, c, patch_size * ratio, patch_size * ratio) ) assert len(dataset_keys) == 4, "length of @dataset_keys should be 4" assert save_path.endswith("h5"), "saved file should end with h5 but get {}".format( save_path.split(".")[-1] ) gt = unfold_fn(torch.tensor(file[dataset_keys[0]][:]), ms_channel, up_ratio) ms = unfold_fn(torch.tensor(file[dataset_keys[1]][:]), ms_channel, 1) lms = unfold_fn(torch.tensor(file[dataset_keys[2]][:]), ms_channel, up_ratio) pan = unfold_fn(torch.tensor(file[dataset_keys[3]][:]), pan_channel, up_ratio) print("clipped datasets shape:") print("{:^20}{:^20}{:^20}{:^20}".format(*[k for k in dataset_keys])) print( "{:^20}{:^20}{:^20}{:^20}".format( str(gt.shape), str(ms.shape), str(lms.shape), str(pan.shape) ) ) base_path = os.path.dirname(save_path) if not os.path.exists(base_path): os.makedirs(base_path) print(f"make dir {base_path}") save_file = h5py.File(save_path, "w") for k, data in zip(dataset_keys, [gt, ms, lms, pan]): save_file.create_dataset(name=k, data=data) print(f"create data {k}") file.close() save_file.close() print("file closed") def dist_gather_object(obj, n_ranks=1, dest=0, all_gather=False): def _iter_tensor_to_rank(rank_obj, dest=0): if isinstance(rank_obj, dict): for k, v in rank_obj.items(): if isinstance(v, torch.Tensor): rank_obj[k] = v.to(dest) elif isinstance(v, Iterable): rank_obj[k] = _iter_tensor_to_rank(v, dest) elif isinstance(rank_obj, (list, tuple)): if isinstance(rank_obj, tuple): rank_obj = list(rank_obj) for i, v in enumerate(rank_obj): if isinstance(v, torch.Tensor): rank_obj[i] = v.to(dest) elif isinstance(v, Iterable): rank_obj[i] = _iter_tensor_to_rank(v, dest) elif isinstance(rank_obj, torch.Tensor): rank_obj = rank_obj.to(dest) return rank_obj if n_ranks == 1: return obj elif n_ranks > 1: rank_objs = [None] * n_ranks if all_gather: # all proc to proc dest dist.all_gather_object(rank_objs, obj) # if is_main_process(): # _scattered_objs_lst = [rank_objs] * n_ranks # else: # _scattered_objs_lst = [None] * n_ranks # received_objs = [None] # dist.scatter_object_list(received_objs, _scattered_objs_lst) rank_objs = _iter_tensor_to_rank(rank_objs, dest=dest) else: dist.gather_object(obj, rank_objs if is_main_process() else None, dest) if is_main_process(): rank_objs = _iter_tensor_to_rank(rank_objs, dest) return rank_objs else: raise ValueError("n_ranks should be greater than 0") if __name__ == "__main__": # path = "/home/ZiHanCao/datasets/HISI/new_harvard/x8/test_harvard(with_up)x8_rgb.h5" # file = h5py.File(path) # clip_dataset_into_small_patches( # file, # patch_size=16, # up_ratio=8, # ms_channel=31, # pan_channel=3, # dataset_keys=["GT", "LRHSI", "HSI_up", "RGB"], # save_path="/home/ZiHanCao/datasets/HISI/new_harvard/x8/test_clip_128.h5", # ) # vis = torch.randn(1, 3, 256, 256).clip(0, 1) # ir = torch.randn(1, 1, 256, 256).clip(0, 1) # model = lambda vis, ir: vis # with y_pred_model_colored(vis, enable=True) as (y, back_to_rgb): # pred_y = model(y, ir) # pred_rgb = back_to_rgb(pred_y) # # assert equal # print(torch.isclose(pred_rgb, vis)) # mean_diff = torch.mean(torch.abs(vis - pred_rgb)) # print("mean difference:", mean_diff.item()) d = dict( a=1, b=2, c=dict( ca=1, cb=2, ) ) args = NameSpace(**d) print(args.a) print(args['c']['ca'])
294coder/Efficient-MIF
3,738
utils/calculate_metric_from_file.py
#%% import numpy as np import torch import matplotlib.pyplot as plt from scipy.io import loadmat from pathlib import Path import h5py from tabulate import tabulate from tqdm import tqdm, trange from utils import AnalysisPanAcc, find_data_path, h5py_to_dict def metric_dicts_ave_and_std(metrics: list[dict]): ave = {} std = {} keys = metrics[0].keys() for k in keys: ave[k] = np.mean([m[k] for m in metrics]) std[k] = np.std([m[k] for m in metrics]) return ave, std def to_tabulate(ave: dict, std: dict): table = [] for k in ave.keys(): table.append([k, ave[k], std[k]]) return tabulate(table, headers=["Metric", "Average", "Standard Deviation"], tablefmt="rounded_grid") def norm_to_0_1(*args, norm_const=2047): return [a / norm_const for a in args] if len(args) > 1 else args[0] / norm_const #%% path = '/volsparse1/czh/exps/fcformer-bk/visualized_img/data_MIMO_SST_harvard_x8_ref.mat' full_res = True if 'unref' in path else False dataset_type = 'harvard_x8' ratio = 4 if dataset_type not in ['cave', 'harvard'] else int(dataset_type.split('_')[-1][-1]) print('===================DATASET========================') print(f'Dataset: {dataset_type}') print(f'Full Resolution: {full_res}') print(f'Ratio: {ratio}') print('=================================================') const = {'wv3': 2047, 'gf': 1023, 'qb': 1023, 'wv2': 2047, 'cave': 1, 'harvard': 1, 'gf5': 1}.get(dataset_type.split('_')[0], 1) dataset_path = find_data_path(dataset_type, full_res) dataset = h5py.File(dataset_path, 'r') if dataset_type[:4] == "cave" or dataset_type[:7] == "harvard": keys = ["LRHSI", "HSI_up", "RGB", "GT"] else: keys = None dataset = h5py_to_dict(dataset, keys) analysis = AnalysisPanAcc(ref=not full_res, ratio=ratio, sensor=dataset_type) path = Path(path) if path.is_dir(): file_lst = list(path.glob('*.mat')) metrics = [] for i, path in tqdm(enumerate(file_lst)): if hasattr(dataset, 'ms'): ms, lms, pan = dataset["ms"][i:i+1], dataset["lms"][i:i+1], dataset["pan"][i:i+1] else: ms, lms, pan = dataset["LRHSI"][i:i+1], dataset["HSI_up"][i:i+1], dataset["RGB"][i:i+1] sr = loadmat(path)['sr'] ms, lms, pan, sr = norm_to_0_1(ms, lms, pan, sr, norm_const=const) if full_res: analysis(sr, ms, lms, pan) else: if hasattr(dataset, 'gt'): gt = dataset["gt"][i:i+1] else: gt = dataset['GT'][i:i+1] gt = norm_to_0_1(gt, norm_const=const) analysis(sr, gt) metrics.append(analysis.acc_ave) analysis.clear_history() ave, std = metric_dicts_ave_and_std(metrics) table = to_tabulate(ave, std) else: files = loadmat(path)['sr'] metrics = [] for i in trange(len(files)): if hasattr(dataset, 'ms'): ms, lms, pan = dataset["ms"][i:i+1], dataset["lms"][i:i+1], dataset["pan"][i:i+1] else: ms, lms, pan = dataset["LRHSI"][i:i+1], dataset["HSI_up"][i:i+1], dataset["RGB"][i:i+1] sr = files[i:i+1] ms, lms, pan, sr = norm_to_0_1(ms, lms, pan, sr, norm_const=const) if full_res: analysis(sr, ms, lms, pan) else: if hasattr(dataset, 'gt'): gt = dataset["gt"][i:i+1] else: gt = dataset['GT'][i:i+1] gt = norm_to_0_1(gt, norm_const=const) analysis(sr, gt) metrics.append(analysis.acc_ave) analysis.clear_history() ave, std = metric_dicts_ave_and_std(metrics) table = to_tabulate(ave, std) print(table)
2977094657/DidaAPI
2,651
frontend/docs/api/auth/validate-wechat-login.md
# 验证微信登录 使用从轮询接口获得的授权码向滴答清单验证微信登录,获取认证令牌。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/user/sign/wechat/validate` - **请求方法**: `GET` - **认证要求**: 无需认证(此接口用于获取认证) - **所属平台**: 滴答清单 ## 请求参数 ### Query Parameters | 参数名 | 类型 | 必填 | 说明 | |--------|------|------|------| | code | string | 是 | 微信授权码,从轮询接口获得 | | state | string | 否 | 状态参数,需与获取二维码时一致 | ## 完整请求示例 ```http GET https://api.dida365.com/api/v2/user/sign/wechat/validate?code={authorization_code}&state={state} HTTP/1.1 Host: api.dida365.com Accept: */* Accept-Encoding: gzip, deflate, br, zstd Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7 Content-Type: application/json Origin: https://dida365.com Referer: https://dida365.com/ User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36 X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":""} ``` ## 响应格式 ### 成功响应 **状态码**: `200 OK` **Content-Type**: `application/json;charset=UTF-8` ### 响应头(重要) 响应头中包含认证相关的Cookie: ```http Set-Cookie: t={auth_token}; Domain=.dida365.com; Expires={expires_date}; Path=/; Secure; HttpOnly; SameSite=None Set-Cookie: _csrf_token={csrf_token}; Domain=.dida365.com; Expires={expires_date}; Path=/; Secure ``` ### 响应体 ```json { "success": true, "message": "string", "user": { "id": "string", "username": "string", "email": "string" } } ``` ## 关键信息提取 ### 认证令牌 (t cookie) 这是最重要的认证信息,用于后续所有API调用: ``` t={auth_token} ``` ### CSRF令牌 (_csrf_token cookie) 用于防止跨站请求伪造攻击: ``` _csrf_token={csrf_token} ``` ## Cookie解析 ### 解析Set-Cookie头 ```python import re def parse_cookies(set_cookie_header): """解析Set-Cookie头""" cookies = {} # 分割多个cookie cookie_parts = set_cookie_header.split(', ') for part in cookie_parts: # 提取cookie名称和值 match = re.match(r'([^=]+)=([^;]+)', part) if match: name = match.group(1).strip() value = match.group(2).strip() # 处理空值cookie(用于删除) if value != '""': cookies[name] = value return cookies ``` ## 后续API调用 获得认证令牌后,所有需要认证的API调用都需要包含: ### 必需的Cookie ```http Cookie: t={认证令牌}; _csrf_token={CSRF令牌} ``` ### 必需的请求头 ```http X-Csrftoken: {CSRF令牌} X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":""} ``` ## 相关接口 - [轮询登录状态](./poll-login-status.md) - 获取用于验证的授权码 - [获取所有任务](../tasks/get-all-tasks.md) - 使用认证令牌获取任务数据
2977094657/DidaAPI
1,427
frontend/docs/api/auth/wechat-callback.md
# 微信登录回调处理 ## 接口信息 - **接口路径**: `GET /auth/wechat/callback` - **接口描述**: 处理微信扫码后的回调,提取code参数 - **请求方式**: GET - **认证要求**: 无需认证 ## 请求参数 ### Query参数 | 参数名 | 类型 | 必填 | 描述 | 示例值 | |--------|------|------|------|--------| | code | string | 是 | 微信返回的授权码 | `001uZIkl2urXGf4qrmml2mAIWz4uZIkW` | | state | string | 否 | 状态参数 | `Lw==` | ## 请求示例 ```bash curl -X GET "http://localhost:8000/auth/wechat/callback?code=001uZIkl2urXGf4qrmml2mAIWz4uZIkW&state=Lw==" ``` ## 响应格式 ### 成功响应 ```json { "message": "微信登录成功", "code": "001uZIkl2urXGf4qrmml2mAIWz4uZIkW", "state": "Lw==", "login_result": { "success": true, "message": "微信登录验证成功", "user_info": { "username": "用户名", "email": "user@example.com" }, "session_info": { "auth_token": "43A001113F9610FFC85AA97B18A297A4...", "csrf_token": "rMwhKGWevLOhHIhFv6hHjAziGDbkpnuY..." } }, "next_step": "用户已成功登录,可以访问受保护的资源" } ``` ### 失败响应 ```json { "message": "微信登录失败", "code": "001uZIkl2urXGf4qrmml2mAIWz4uZIkW", "state": "Lw==", "error": "验证失败" } ``` ## 使用说明 1. **回调处理**: 这个接口用于演示微信回调的处理流程 2. **自动验证**: 接收到回调后会自动进行登录验证 3. **会话建立**: 验证成功后会建立用户会话 4. **实际使用**: 在实际使用中,微信会重定向到配置的redirect_uri ## 相关接口 - [微信登录流程](./wechat-login-flow.md) - [获取微信二维码](./get-wechat-qrcode.md) - [轮询登录状态](./poll-login-status.md) - [验证微信登录](./validate-wechat-login.md) ## 注意事项 - 这是微信OAuth流程的最后一步 - code参数只能使用一次,有效期很短 - 成功后会自动建立认证会话 - 可以直接用于后续API调用
2977094657/DidaAPI
1,931
frontend/docs/api/auth/poll-login-status.md
# 轮询微信登录状态 通过长轮询检查用户是否已扫码并确认微信登录。 ## 接口信息 - **接口URL**: `https://long.open.weixin.qq.com/connect/l/qrconnect` - **请求方法**: `GET` - **认证要求**: 无需认证 - **所属平台**: 微信开放平台 ## 请求参数 ### Query Parameters | 参数名 | 类型 | 必填 | 说明 | |--------|------|------|------| | uuid | string | 是 | 二维码密钥,从获取二维码接口中提取的16位字符串 | | _ | number | 是 | 时间戳,用于防止缓存,格式:毫秒级时间戳 | ## 完整请求示例 ```http GET https://long.open.weixin.qq.com/connect/l/qrconnect?uuid={qr_code_key}&_={timestamp} HTTP/1.1 Host: long.open.weixin.qq.com User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36 Accept: */* Accept-Language: zh-CN,zh;q=0.9,en;q=0.8 Referer: https://open.weixin.qq.com/ ``` ## 响应格式 ### 响应类型 **Content-Type**: `text/javascript; charset=utf-8` **响应体**: JavaScript代码片段 ### 状态码说明 响应中包含 `window.wx_errcode` 表示不同的状态: | errcode | 说明 | 后续操作 | |---------|------|----------| | 404 | 等待扫码 | 继续轮询 | | 403 | 已扫码,等待用户确认 | 继续轮询 | | 405 | 登录成功,获得授权码 | 停止轮询,提取授权码 | | 408 | 二维码已过期 | 停止轮询,重新获取二维码 | | 400 | 二维码已失效 | 停止轮询,重新获取二维码 | ## 响应示例 ### 等待扫码 (errcode=404) ```javascript window.wx_errcode=404;window.wx_code=''; ``` ### 已扫码等待确认 (errcode=403) ```javascript window.wx_errcode=403;window.wx_code=''; ``` ### 登录成功 (errcode=405) ```javascript window.wx_errcode=405;window.wx_code='{authorization_code}'; ``` ### 二维码过期 (errcode=408) ```javascript window.wx_errcode=408;window.wx_code=''; ``` ## 响应解析 ### 提取状态码和授权码 ```javascript // 解析响应文本 function parseWeChatResponse(responseText) { const errcodeMatch = responseText.match(/window\.wx_errcode\s*=\s*(\d+)/); const codeMatch = responseText.match(/window\.wx_code\s*=\s*'([^']*)'/); return { errcode: errcodeMatch ? parseInt(errcodeMatch[1]) : null, code: codeMatch ? codeMatch[1] : '' }; } ``` ## 相关接口 - [获取微信二维码](./get-wechat-qrcode.md) - 获取用于轮询的二维码密钥 - [验证微信登录](./validate-wechat-login.md) - 使用获得的授权码进行登录验证
2977094657/DidaAPI
2,177
frontend/docs/api/auth/get-wechat-qrcode.md
# 获取微信登录二维码 获取微信扫码登录的二维码,这是微信登录流程的第一步。 ## 接口信息 - **接口URL**: `https://open.weixin.qq.com/connect/qrconnect` - **请求方法**: `GET` - **认证要求**: 无需认证 - **所属平台**: 微信开放平台 ## 请求参数 ### Query Parameters | 参数名 | 类型 | 必填 | 说明 | |--------|------|------|------| | appid | string | 是 | 微信应用ID | | redirect_uri | string | 是 | 授权后重定向的回调地址 | | response_type | string | 是 | 返回类型,固定值:`code` | | scope | string | 是 | 应用授权作用域,固定值:`snsapi_login` | | state | string | 否 | 重定向后会带上state参数,开发者可以填写a-zA-Z0-9的参数值,最多128字节 | ## 完整请求示例 ```http GET https://open.weixin.qq.com/connect/qrconnect?appid={appid}&redirect_uri={redirect_uri}&response_type=code&scope=snsapi_login&state={state} HTTP/1.1 Host: open.weixin.qq.com User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36 Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8 Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7 Accept-Encoding: gzip, deflate, br, zstd ``` ## 响应格式 ### 成功响应 **状态码**: `200 OK` **Content-Type**: `text/html; charset=utf-8` **响应体**: HTML页面,包含二维码图片 ### 关键HTML结构 ```html <!DOCTYPE html> <html> <head> <title>微信登录</title> </head> <body> <!-- 其他HTML内容 --> <img class="qrcode lightBorder js_qrcode_img" src="/connect/qrcode/{qr_code_key}"> <!-- 其他HTML内容 --> </body> </html> ``` ## 响应解析 ### 提取二维码密钥 从响应HTML中提取二维码图片的src属性: ```javascript // 正则表达式匹配 const pattern = /<img[^>]*class="[^"]*qrcode[^"]*"[^>]*src="([^"]*)"/; const match = html.match(pattern); if (match) { const srcUrl = match[1]; // "/connect/qrcode/{qr_code_key}" const qrCodeKey = srcUrl.split('/').pop(); // "{qr_code_key}" } ``` ### 构建完整二维码URL ```javascript const qrCodeUrl = `https://open.weixin.qq.com${srcUrl}`; // 结果: https://open.weixin.qq.com/connect/qrcode/{qr_code_key} ``` ## 响应头示例 ```http HTTP/1.1 200 OK Content-Type: text/html; charset=utf-8 Cache-Control: no-cache, must-revalidate Content-Encoding: gzip Content-Length: 17543 X-Wx-Fj: 001,018,0000001024 ``` ## 后续步骤 获取二维码后,需要: 1. 显示二维码供用户扫描 2. 使用二维码密钥进行[轮询登录状态](./poll-login-status.md) 3. 获取授权码后进行[验证微信登录](./validate-wechat-login.md)
2977094657/DidaAPI
1,914
frontend/docs/api/auth/wechat-login-flow.md
# 微信扫码登录完整流程 微信扫码登录到滴答清单是一个多步骤的过程,涉及微信开放平台和滴答清单API的协作。 ## 完整流程概览 ```mermaid sequenceDiagram participant User as 用户 participant Client as 客户端 participant WeChat as 微信服务器 participant Dida as 滴答清单服务器 User->>Client: 1. 请求登录 Client->>WeChat: 2. 获取二维码 WeChat-->>Client: 3. 返回二维码HTML Client->>Client: 4. 解析二维码URL Client->>User: 5. 显示二维码 loop 轮询检查 Client->>WeChat: 6. 轮询登录状态 WeChat-->>Client: 7. 返回状态 end User->>WeChat: 8. 微信扫码确认 WeChat-->>Client: 9. 返回授权码 Client->>Dida: 10. 验证授权码 Dida-->>Client: 11. 返回认证令牌 Client->>Client: 12. 保存认证信息 ``` ## 详细步骤 ### 步骤1: 获取微信二维码 **接口**: [获取微信二维码](./get-wechat-qrcode.md) 请求微信开放平台获取登录二维码: ```http GET https://open.weixin.qq.com/connect/qrconnect?appid=wxf1429a73d311aad4&redirect_uri=https://dida365.com/sign/wechat&response_type=code&scope=snsapi_login&state=Lw== ``` ### 步骤2: 解析二维码信息 从返回的HTML中提取二维码图片URL: ```html <img class="qrcode lightBorder js_qrcode_img" src="/connect/qrcode/071cfvEZ3xaPll2J"> ``` 提取16位密钥:`071cfvEZ3xaPll2J` ### 步骤3: 轮询登录状态 **接口**: [轮询登录状态](./poll-login-status.md) 使用长轮询检查用户是否扫码: ```http GET https://long.open.weixin.qq.com/connect/l/qrconnect?uuid=071cfvEZ3xaPll2J&_=1748955598000 ``` ### 步骤4: 用户扫码确认 用户使用微信扫描二维码并确认登录。 ### 步骤5: 获取授权码 轮询接口返回授权码: ```javascript window.wx_errcode=405;window.wx_code='001uZIkl2urXGf4qrmml2mAIWz4uZIkW'; ``` ### 步骤6: 验证登录 **接口**: [验证微信登录](./validate-wechat-login.md) 使用授权码向滴答清单验证登录: ```http GET https://api.dida365.com/api/v2/user/sign/wechat/validate?code=001uZIkl2urXGf4qrmml2mAIWz4uZIkW&state=Lw== ``` ### 步骤7: 获取认证令牌 验证成功后获得认证令牌和CSRF令牌: ```http Set-Cookie: t=43A001113F9610FFC85AA97B18A297A4...; Domain=.dida365.com Set-Cookie: _csrf_token=rMwhKGWevLOhHIhFv6hHjAziGDbkpnuY...; Domain=.dida365.com ``` ## 相关接口 - [获取微信二维码](./get-wechat-qrcode.md) - [轮询登录状态](./poll-login-status.md) - [验证微信登录](./validate-wechat-login.md)
281677160/openwrt-package
94,932
luci-app-passwall/root/usr/share/passwall/app.sh
#!/bin/sh # Copyright (C) 2018-2020 L-WRT Team # Copyright (C) 2021-2025 xiaorouji . $IPKG_INSTROOT/lib/functions.sh . $IPKG_INSTROOT/lib/functions/service.sh CONFIG=passwall TMP_PATH=/tmp/etc/$CONFIG TMP_BIN_PATH=$TMP_PATH/bin TMP_SCRIPT_FUNC_PATH=$TMP_PATH/script_func TMP_ROUTE_PATH=$TMP_PATH/route TMP_ACL_PATH=$TMP_PATH/acl TMP_IFACE_PATH=$TMP_PATH/iface TMP_PATH2=/tmp/etc/${CONFIG}_tmp GLOBAL_ACL_PATH=${TMP_ACL_PATH}/default LOG_FILE=/tmp/log/$CONFIG.log APP_PATH=/usr/share/$CONFIG RULES_PATH=/usr/share/${CONFIG}/rules LUA_UTIL_PATH=/usr/lib/lua/luci/passwall UTIL_SINGBOX=$LUA_UTIL_PATH/util_sing-box.lua UTIL_SS=$LUA_UTIL_PATH/util_shadowsocks.lua UTIL_XRAY=$LUA_UTIL_PATH/util_xray.lua UTIL_TROJAN=$LUA_UTIL_PATH/util_trojan.lua UTIL_NAIVE=$LUA_UTIL_PATH/util_naiveproxy.lua UTIL_HYSTERIA2=$LUA_UTIL_PATH/util_hysteria2.lua UTIL_TUIC=$LUA_UTIL_PATH/util_tuic.lua echolog() { local d="$(date "+%Y-%m-%d %H:%M:%S")" echo -e "$d: $*" >>$LOG_FILE } config_get_type() { local ret=$(uci -q get "${CONFIG}.${1}" 2>/dev/null) echo "${ret:=$2}" } config_n_get() { local ret=$(uci -q get "${CONFIG}.${1}.${2}" 2>/dev/null) echo "${ret:=$3}" } config_t_get() { local index=${4:-0} local ret=$(uci -q get "${CONFIG}.@${1}[${index}].${2}" 2>/dev/null) echo "${ret:=${3}}" } config_t_set() { local index=${4:-0} local ret=$(uci -q set "${CONFIG}.@${1}[${index}].${2}=${3}" 2>/dev/null) } get_enabled_anonymous_secs() { uci -q show "${CONFIG}" | grep "${1}\[.*\.enabled='1'" | cut -d '.' -sf2 } get_host_ip() { local host=$2 local count=$3 [ -z "$count" ] && count=3 local isip="" local ip=$host if [ "$1" == "ipv6" ]; then isip=$(echo $host | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}") if [ -n "$isip" ]; then isip=$(echo $host | cut -d '[' -f2 | cut -d ']' -f1) fi else isip=$(echo $host | grep -E "([0-9]{1,3}[\.]){3}[0-9]{1,3}") fi [ -z "$isip" ] && { local t=4 [ "$1" == "ipv6" ] && t=6 local vpsrip=$(resolveip -$t -t $count $host | awk 'NR==1{print}') ip=$vpsrip } echo $ip } get_node_host_ip() { local ip local address=$(config_n_get $1 address) [ -n "$address" ] && { local use_ipv6=$(config_n_get $1 use_ipv6) local network_type="ipv4" [ "$use_ipv6" == "1" ] && network_type="ipv6" ip=$(get_host_ip $network_type $address) } echo $ip } get_ip_port_from() { local __host=${1}; shift 1 local __ipv=${1}; shift 1 local __portv=${1}; shift 1 local __ucipriority=${1}; shift 1 local val1 val2 if [ -n "${__ucipriority}" ]; then val2=$(config_n_get ${__host} port $(echo $__host | sed -n 's/^.*[:#]\([0-9]*\)$/\1/p')) val1=$(config_n_get ${__host} address "${__host%%${val2:+[:#]${val2}*}}") else val2=$(echo $__host | sed -n 's/^.*[:#]\([0-9]*\)$/\1/p') val1="${__host%%${val2:+[:#]${val2}*}}" fi eval "${__ipv}=\"$val1\"; ${__portv}=\"$val2\"" } host_from_url(){ local f=${1} ## Remove protocol part of url ## f="${f##http://}" f="${f##https://}" f="${f##ftp://}" f="${f##sftp://}" ## Remove username and/or username:password part of URL ## f="${f##*:*@}" f="${f##*@}" ## Remove rest of urls ## f="${f%%/*}" echo "${f%%:*}" } hosts_foreach() { local __hosts eval "__hosts=\$${1}"; shift 1 local __func=${1}; shift 1 local __default_port=${1}; shift 1 local __ret=1 [ -z "${__hosts}" ] && return 0 local __ip __port for __host in $(echo $__hosts | sed 's/[ ,]/\n/g'); do get_ip_port_from "$__host" "__ip" "__port" eval "$__func \"${__host}\" \"\${__ip}\" \"\${__port:-${__default_port}}\" \"$@\"" __ret=$? [ ${__ret} -ge ${ERROR_NO_CATCH:-1} ] && return ${__ret} done } check_host() { local f=${1} a=$(echo $f | grep "\/") [ -n "$a" ] && return 1 # 判断是否包含汉字~ local tmp=$(echo -n $f | awk '{print gensub(/[!-~]/,"","g",$0)}') [ -n "$tmp" ] && return 1 return 0 } get_first_dns() { local __hosts_val=${1}; shift 1 __first() { [ -z "${2}" ] && return 0 echo "${2}#${3}" return 1 } eval "hosts_foreach \"${__hosts_val}\" __first \"$@\"" } get_last_dns() { local __hosts_val=${1}; shift 1 local __first __last __every() { [ -z "${2}" ] && return 0 __last="${2}#${3}" __first=${__first:-${__last}} } eval "hosts_foreach \"${__hosts_val}\" __every \"$@\"" [ "${__first}" == "${__last}" ] || echo "${__last}" } check_port_exists() { local port=$1 local protocol=$2 [ -n "$protocol" ] || protocol="tcp,udp" local result= if [ "$protocol" = "tcp" ]; then result=$(netstat -tln | grep -c ":$port ") elif [ "$protocol" = "udp" ]; then result=$(netstat -uln | grep -c ":$port ") elif [ "$protocol" = "tcp,udp" ]; then result=$(netstat -tuln | grep -c ":$port ") fi echo "${result}" } get_new_port() { local port=$1 [ "$port" == "auto" ] && port=2082 local protocol=$(echo $2 | tr 'A-Z' 'a-z') local result=$(check_port_exists $port $protocol) if [ "$result" != 0 ]; then local temp= if [ "$port" -lt 65535 ]; then temp=$(expr $port + 1) elif [ "$port" -gt 1 ]; then temp=$(expr $port - 1) fi get_new_port $temp $protocol else echo $port fi } check_depends() { local depends local tables=${1} local file_path="/usr/lib/opkg/info" local file_ext=".control" [ -d "/lib/apk/packages" ] && file_path="/lib/apk/packages" && file_ext=".list" if [ "$tables" == "iptables" ]; then for depends in "iptables-mod-tproxy" "iptables-mod-socket" "iptables-mod-iprange" "iptables-mod-conntrack-extra" "kmod-ipt-nat"; do [ -s "${file_path}/${depends}${file_ext}" ] || echolog "$tables透明代理基础依赖 $depends 未安装..." done else for depends in "kmod-nft-socket" "kmod-nft-tproxy" "kmod-nft-nat"; do [ -s "${file_path}/${depends}${file_ext}" ] || echolog "$tables透明代理基础依赖 $depends 未安装..." done fi } check_ver() { local version1="$1" local version2="$2" local i v1 v1_1 v1_2 v1_3 v2 v2_1 v2_2 v2_3 IFS='.'; set -- $version1; v1_1=${1:-0}; v1_2=${2:-0}; v1_3=${3:-0} IFS='.'; set -- $version2; v2_1=${1:-0}; v2_2=${2:-0}; v2_3=${3:-0} IFS= for i in 1 2 3; do eval v1=\$v1_$i eval v2=\$v2_$i if [ "$v1" -gt "$v2" ]; then # $1 大于 $2 echo 0 return elif [ "$v1" -lt "$v2" ]; then # $1 小于 $2 echo 1 return fi done # $1 等于 $2 echo 255 } first_type() { for p in "/bin/$1" "${TMP_BIN_PATH:-/tmp}/$1" "$1"; do [ -x "$p" ] && echo "$p" && return done command -v "$1" 2>/dev/null || command -v "$2" 2>/dev/null } eval_set_val() { for i in $@; do for j in $i; do eval $j done done } eval_unset_val() { for i in $@; do for j in $i; do eval unset j done done } ln_run() { local file_func=${1} local ln_name=${2} local output=${3} shift 3; if [ "${file_func%%/*}" != "${file_func}" ]; then [ ! -L "${file_func}" ] && { ln -s "${file_func}" "${TMP_BIN_PATH}/${ln_name}" >/dev/null 2>&1 file_func="${TMP_BIN_PATH}/${ln_name}" } [ -x "${file_func}" ] || echolog " - $(readlink ${file_func}) 没有执行权限,无法启动:${file_func} $*" fi #echo "${file_func} $*" >&2 [ -n "${file_func}" ] || echolog " - 找不到 ${ln_name},无法启动..." [ "${output}" != "/dev/null" ] && [ "${ln_name}" != "chinadns-ng" ] && { local persist_log_path=$(config_t_get global persist_log_path) local sys_log=$(config_t_get global sys_log "0") } if [ -z "$persist_log_path" ] && [ "$sys_log" != "1" ]; then ${file_func:-echolog " - ${ln_name}"} "$@" >${output} 2>&1 & else [ "${output: -1, -7}" == "TCP.log" ] && local protocol="TCP" [ "${output: -1, -7}" == "UDP.log" ] && local protocol="UDP" if [ -n "${persist_log_path}" ]; then mkdir -p ${persist_log_path} local log_file=${persist_log_path}/passwall_${protocol}_${ln_name}_$(date '+%F').log echolog "记录到持久性日志文件:${log_file}" ${file_func:-echolog " - ${ln_name}"} "$@" >> ${log_file} 2>&1 & sys_log=0 fi if [ "${sys_log}" == "1" ]; then echolog "记录 ${ln_name}_${protocol} 到系统日志" ${file_func:-echolog " - ${ln_name}"} "$@" 2>&1 | logger -t PASSWALL_${protocol}_${ln_name} & fi fi process_count=$(ls $TMP_SCRIPT_FUNC_PATH | wc -l) process_count=$((process_count + 1)) echo "${file_func:-echolog " - ${ln_name}"} $@ >${output}" > $TMP_SCRIPT_FUNC_PATH/$process_count } lua_api() { local func=${1} [ -z "${func}" ] && { echo "nil" return } echo $(lua -e "local api = require 'luci.passwall.api' print(api.${func})") } parse_doh() { local __doh=$1 __url_var=$2 __host_var=$3 __port_var=$4 __bootstrap_var=$5 __doh=$(echo -e "$__doh" | tr -d ' \t\n') local __url=${__doh%%,*} local __bootstrap=${__doh#*,} local __host_port=$(lua_api "get_domain_from_url(\"${__url}\")") local __host __port if echo "${__host_port}" | grep -q '^\[.*\]:[0-9]\+$'; then __host=${__host_port%%]:*}] __port=${__host_port##*:} elif echo "${__host_port}" | grep -q ':[0-9]\+$'; then __host=${__host_port%:*} __port=${__host_port##*:} else __host=${__host_port} __port=443 fi __host=${__host#[} __host=${__host%]} if [ "$(lua_api "is_ip(\"${__host}\")")" = "true" ]; then __bootstrap=${__host} fi __bootstrap=${__bootstrap#[} __bootstrap=${__bootstrap%]} eval "${__url_var}='${__url}' ${__host_var}='${__host}' ${__port_var}='${__port}' ${__bootstrap_var}='${__bootstrap}'" } get_geoip() { local geoip_code="$1" local geoip_type_flag="" local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat" [ -s "$geoip_path" ] || { echo ""; return 1; } case "$2" in "ipv4") geoip_type_flag="-ipv6=false" ;; "ipv6") geoip_type_flag="-ipv4=false" ;; esac if type geoview &> /dev/null; then geoview -input "$geoip_path" -list "$geoip_code" $geoip_type_flag -lowmem=true return 0 else echo "" return 1 fi } set_cache_var() { local key="${1}" shift 1 local val="$@" [ -n "${key}" ] && [ -n "${val}" ] && { sed -i "/${key}=/d" $TMP_PATH/var >/dev/null 2>&1 echo "${key}=\"${val}\"" >> $TMP_PATH/var eval ${key}=\"${val}\" } } get_cache_var() { local key="${1}" [ -n "${key}" ] && [ -s "$TMP_PATH/var" ] && { echo $(cat $TMP_PATH/var | grep "^${key}=" | awk -F '=' '{print $2}' | tail -n 1 | awk -F'"' '{print $2}') } } eval_cache_var() { [ -s "$TMP_PATH/var" ] && eval $(cat "$TMP_PATH/var") } has_1_65535() { local val="$1" val=${val//:/-} case ",$val," in *,1-65535,*) return 0 ;; *) return 1 ;; esac } run_ipt2socks() { local flag proto tcp_tproxy local_port socks_address socks_port socks_username socks_password log_file local _extra_param="" eval_set_val $@ [ -n "$log_file" ] || log_file="/dev/null" socks_address=$(get_host_ip "ipv4" ${socks_address}) [ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} -a $socks_username -k $socks_password" [ -n "$tcp_tproxy" ] || _extra_param="${_extra_param} -R" case "$proto" in UDP) flag="${flag}_UDP" _extra_param="${_extra_param} -U" ;; TCP) flag="${flag}_TCP" _extra_param="${_extra_param} -T" ;; *) flag="${flag}_TCP_UDP" ;; esac _extra_param="${_extra_param} -n 65535 -v" ln_run "$(first_type ipt2socks)" "ipt2socks_${flag}" $log_file -l $local_port -b 0.0.0.0 -s $socks_address -p $socks_port ${_extra_param} } run_singbox() { local flag type node tcp_redir_port tcp_proxy_way udp_redir_port socks_address socks_port socks_username socks_password http_address http_port http_username http_password local dns_listen_port direct_dns_query_strategy direct_dns_port direct_dns_udp_server direct_dns_tcp_server remote_dns_protocol remote_dns_udp_server remote_dns_tcp_server remote_dns_doh remote_dns_client_ip remote_fakedns remote_dns_query_strategy dns_cache dns_socks_address dns_socks_port local loglevel log_file config_file server_host server_port no_run local _extra_param="" eval_set_val $@ [ -z "$type" ] && { local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z') if [ "$type" != "sing-box" ]; then bin=$(first_type $(config_t_get global_app sing_box_file) sing-box) [ -n "$bin" ] && type="sing-box" fi } [ -z "$type" ] && return 1 [ -n "$log_file" ] || local log_file="/dev/null" _extra_param="${_extra_param} -log 1 -logfile ${log_file}" if [ "$log_file" = "/dev/null" ]; then _extra_param="${_extra_param} -log 0" else _extra_param="${_extra_param} -log 1 -logfile ${log_file}" fi [ -z "$loglevel" ] && local loglevel=$(config_t_get global loglevel "warn") [ "$loglevel" = "warning" ] && loglevel="warn" _extra_param="${_extra_param} -loglevel $loglevel" [ -n "$flag" ] && _extra_param="${_extra_param} -flag $flag" [ -n "$node" ] && _extra_param="${_extra_param} -node $node" [ -n "$server_host" ] && _extra_param="${_extra_param} -server_host $server_host" [ -n "$server_port" ] && _extra_param="${_extra_param} -server_port $server_port" [ -n "$tcp_redir_port" ] && _extra_param="${_extra_param} -tcp_redir_port $tcp_redir_port" [ -n "$tcp_proxy_way" ] && _extra_param="${_extra_param} -tcp_proxy_way $tcp_proxy_way" [ -n "$udp_redir_port" ] && _extra_param="${_extra_param} -udp_redir_port $udp_redir_port" [ -n "$socks_address" ] && _extra_param="${_extra_param} -local_socks_address $socks_address" [ -n "$socks_port" ] && _extra_param="${_extra_param} -local_socks_port $socks_port" [ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} -local_socks_username $socks_username -local_socks_password $socks_password" [ -n "$http_address" ] && _extra_param="${_extra_param} -local_http_address $http_address" [ -n "$http_port" ] && _extra_param="${_extra_param} -local_http_port $http_port" [ -n "$http_username" ] && [ -n "$http_password" ] && _extra_param="${_extra_param} -local_http_username $http_username -local_http_password $http_password" [ -n "$dns_socks_address" ] && [ -n "$dns_socks_port" ] && _extra_param="${_extra_param} -dns_socks_address ${dns_socks_address} -dns_socks_port ${dns_socks_port}" [ -n "$dns_listen_port" ] && _extra_param="${_extra_param} -dns_listen_port ${dns_listen_port}" [ -n "$dns_cache" ] && _extra_param="${_extra_param} -dns_cache ${dns_cache}" if [ -n "$direct_dns_udp_server" ]; then direct_dns_port=$(echo ${direct_dns_udp_server} | awk -F '#' '{print $2}') _extra_param="${_extra_param} -direct_dns_udp_server $(echo ${direct_dns_udp_server} | awk -F '#' '{print $1}')" elif [ -n "$direct_dns_tcp_server" ]; then direct_dns_port=$(echo ${direct_dns_tcp_server} | awk -F '#' '{print $2}') _extra_param="${_extra_param} -direct_dns_tcp_server $(echo ${direct_dns_tcp_server} | awk -F '#' '{print $1}')" else local local_dns=$(echo -n $(echo "${LOCAL_DNS}" | sed "s/,/\n/g" | head -n1) | tr " " ",") _extra_param="${_extra_param} -direct_dns_udp_server $(echo ${local_dns} | awk -F '#' '{print $1}')" direct_dns_port=$(echo ${local_dns} | awk -F '#' '{print $2}') fi _extra_param="${_extra_param} -direct_dns_port ${direct_dns_port:-53}" direct_dns_query_strategy=${direct_dns_query_strategy:-UseIP} _extra_param="${_extra_param} -direct_dns_query_strategy ${direct_dns_query_strategy}" [ -n "$remote_dns_query_strategy" ] && _extra_param="${_extra_param} -remote_dns_query_strategy ${remote_dns_query_strategy}" case "$remote_dns_protocol" in tcp) local _dns=$(get_first_dns remote_dns_tcp_server 53 | sed 's/#/:/g') local _dns_address=$(echo ${_dns} | awk -F ':' '{print $1}') local _dns_port=$(echo ${_dns} | awk -F ':' '{print $2}') _extra_param="${_extra_param} -remote_dns_server ${_dns_address} -remote_dns_port ${_dns_port} -remote_dns_tcp_server tcp://${_dns}" ;; doh) local _doh_url _doh_host _doh_port _doh_bootstrap parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap [ -n "$_doh_bootstrap" ] && _extra_param="${_extra_param} -remote_dns_server ${_doh_bootstrap}" _extra_param="${_extra_param} -remote_dns_port ${_doh_port} -remote_dns_doh_url ${_doh_url} -remote_dns_doh_host ${_doh_host}" ;; esac [ -n "$remote_dns_client_ip" ] && _extra_param="${_extra_param} -remote_dns_client_ip ${remote_dns_client_ip}" [ "$remote_fakedns" = "1" ] && _extra_param="${_extra_param} -remote_dns_fake 1" [ -n "$no_run" ] && _extra_param="${_extra_param} -no_run 1" lua $UTIL_SINGBOX gen_config ${_extra_param} > $config_file [ -n "$no_run" ] || ln_run "$(first_type $(config_t_get global_app sing_box_file) sing-box)" "sing-box" $log_file run -c "$config_file" } run_xray() { local flag type node tcp_redir_port tcp_proxy_way udp_redir_port socks_address socks_port socks_username socks_password http_address http_port http_username http_password local dns_listen_port direct_dns_query_strategy direct_dns_port direct_dns_udp_server direct_dns_tcp_server remote_dns_udp_server remote_dns_tcp_server remote_dns_doh remote_dns_client_ip remote_fakedns remote_dns_query_strategy dns_cache dns_socks_address dns_socks_port local loglevel log_file config_file server_host server_port no_run local _extra_param="" eval_set_val $@ [ -z "$type" ] && { local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z') if [ "$type" != "xray" ]; then bin=$(first_type $(config_t_get global_app xray_file) xray) [ -n "$bin" ] && type="xray" fi } [ -z "$type" ] && return 1 [ -n "$log_file" ] || local log_file="/dev/null" [ -z "$loglevel" ] && local loglevel=$(config_t_get global loglevel "warning") [ -n "$flag" ] && _extra_param="${_extra_param} -flag $flag" [ -n "$node" ] && _extra_param="${_extra_param} -node $node" [ -n "$server_host" ] && _extra_param="${_extra_param} -server_host $server_host" [ -n "$server_port" ] && _extra_param="${_extra_param} -server_port $server_port" [ -n "$tcp_redir_port" ] && _extra_param="${_extra_param} -tcp_redir_port $tcp_redir_port" [ -n "$tcp_proxy_way" ] && _extra_param="${_extra_param} -tcp_proxy_way $tcp_proxy_way" [ -n "$udp_redir_port" ] && _extra_param="${_extra_param} -udp_redir_port $udp_redir_port" [ -n "$socks_address" ] && _extra_param="${_extra_param} -local_socks_address $socks_address" [ -n "$socks_port" ] && _extra_param="${_extra_param} -local_socks_port $socks_port" [ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} -local_socks_username $socks_username -local_socks_password $socks_password" [ -n "$http_address" ] && _extra_param="${_extra_param} -local_http_address $http_address" [ -n "$http_port" ] && _extra_param="${_extra_param} -local_http_port $http_port" [ -n "$http_username" ] && [ -n "$http_password" ] && _extra_param="${_extra_param} -local_http_username $http_username -local_http_password $http_password" [ -n "$dns_socks_address" ] && [ -n "$dns_socks_port" ] && _extra_param="${_extra_param} -dns_socks_address ${dns_socks_address} -dns_socks_port ${dns_socks_port}" [ -n "$dns_listen_port" ] && _extra_param="${_extra_param} -dns_listen_port ${dns_listen_port}" if [ -n "$direct_dns_udp_server" ]; then direct_dns_port=$(echo ${direct_dns_udp_server} | awk -F '#' '{print $2}') _extra_param="${_extra_param} -direct_dns_udp_server $(echo ${direct_dns_udp_server} | awk -F '#' '{print $1}')" elif [ -n "$direct_dns_tcp_server" ]; then direct_dns_port=$(echo ${direct_dns_tcp_server} | awk -F '#' '{print $2}') _extra_param="${_extra_param} -direct_dns_tcp_server $(echo ${direct_dns_tcp_server} | awk -F '#' '{print $1}')" else local local_dns=$(echo -n $(echo "${LOCAL_DNS}" | sed "s/,/\n/g" | head -n1) | tr " " ",") _extra_param="${_extra_param} -direct_dns_udp_server $(echo ${local_dns} | awk -F '#' '{print $1}')" direct_dns_port=$(echo ${local_dns} | awk -F '#' '{print $2}') fi _extra_param="${_extra_param} -direct_dns_port ${direct_dns_port:-53}" direct_dns_query_strategy=${direct_dns_query_strategy:-UseIP} _extra_param="${_extra_param} -direct_dns_query_strategy ${direct_dns_query_strategy}" [ -n "$remote_dns_query_strategy" ] && _extra_param="${_extra_param} -remote_dns_query_strategy ${remote_dns_query_strategy}" [ -n "$remote_dns_client_ip" ] && _extra_param="${_extra_param} -remote_dns_client_ip ${remote_dns_client_ip}" [ "$remote_fakedns" = "1" ] && _extra_param="${_extra_param} -remote_dns_fake 1" [ -n "$dns_cache" ] && _extra_param="${_extra_param} -dns_cache ${dns_cache}" [ -n "${remote_dns_tcp_server}" ] && { local _dns=$(get_first_dns remote_dns_tcp_server 53 | sed 's/#/:/g') local _dns_address=$(echo ${_dns} | awk -F ':' '{print $1}') local _dns_port=$(echo ${_dns} | awk -F ':' '{print $2}') _extra_param="${_extra_param} -remote_dns_tcp_server ${_dns_address} -remote_dns_tcp_port ${_dns_port}" } [ -n "${remote_dns_doh}" ] && { local _doh_url _doh_host _doh_port _doh_bootstrap parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap [ -n "$_doh_bootstrap" ] && _extra_param="${_extra_param} -remote_dns_doh_ip ${_doh_bootstrap}" _extra_param="${_extra_param} -remote_dns_doh_port ${_doh_port} -remote_dns_doh_url ${_doh_url} -remote_dns_doh_host ${_doh_host}" } _extra_param="${_extra_param} -loglevel $loglevel" [ -n "$no_run" ] && _extra_param="${_extra_param} -no_run 1" lua $UTIL_XRAY gen_config ${_extra_param} > $config_file [ -n "$no_run" ] || ln_run "$(first_type $(config_t_get global_app ${type}_file) ${type})" ${type} $log_file run -c "$config_file" } run_dns2socks() { local flag socks socks_address socks_port socks_username socks_password listen_address listen_port dns cache log_file local _extra_param="" eval_set_val $@ [ -n "$flag" ] && flag="_${flag}" [ -n "$log_file" ] || log_file="/dev/null" dns=$(get_first_dns dns 53 | sed 's/#/:/g') [ -n "$socks" ] && { socks="${socks//#/:}" socks_address=$(echo $socks | awk -F ':' '{print $1}') socks_port=$(echo $socks | awk -F ':' '{print $2}') } [ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} /u $socks_username /p $socks_password" [ -z "$cache" ] && cache=1 [ "$cache" = "0" ] && _extra_param="${_extra_param} /d" ln_run "$(first_type dns2socks)" "dns2socks${flag}" $log_file ${_extra_param} "${socks_address}:${socks_port}" "${dns}" "${listen_address}:${listen_port}" } run_chinadns_ng() { local _flag _listen_port _dns_local _dns_trust _no_ipv6_trust _use_direct_list _use_proxy_list _gfwlist _chnlist _default_mode _default_tag _no_logic_log _tcp_node _remote_fakedns local _extra_param="" eval_set_val $@ local _CONF_FILE=$TMP_ACL_PATH/$_flag/chinadns_ng.conf local _LOG_FILE="/dev/null" [ "$(config_t_get global log_chinadns_ng "0")" == "1" ] && _LOG_FILE=$TMP_ACL_PATH/$_flag/chinadns_ng.log _extra_param="-FLAG ${_flag} -TCP_NODE ${_tcp_node} -LISTEN_PORT ${_listen_port} -DNS_LOCAL ${_dns_local} -DNS_TRUST ${_dns_trust}" _extra_param="${_extra_param} -USE_DIRECT_LIST ${_use_direct_list} -USE_PROXY_LIST ${_use_proxy_list} -USE_BLOCK_LIST ${_use_block_list}" _extra_param="${_extra_param} -GFWLIST ${_gfwlist} -CHNLIST ${_chnlist} -NO_IPV6_TRUST ${_no_ipv6_trust} -DEFAULT_MODE ${_default_mode}" _extra_param="${_extra_param} -DEFAULT_TAG ${_default_tag} -NFTFLAG ${nftflag} -NO_LOGIC_LOG ${_no_logic_log} -REMOTE_FAKEDNS ${_remote_fakedns}" _extra_param="${_extra_param} -LOG_FILE ${_LOG_FILE}" lua $APP_PATH/helper_chinadns_add.lua ${_extra_param} > ${_CONF_FILE} ln_run "$(first_type chinadns-ng)" chinadns-ng "${_LOG_FILE}" -C ${_CONF_FILE} } run_socks() { local flag node bind socks_port config_file http_port http_config_file relay_port log_file no_run eval_set_val $@ [ -n "$config_file" ] && [ -z "$(echo ${config_file} | grep $TMP_PATH)" ] && config_file=$TMP_PATH/$config_file [ -n "$http_port" ] || http_port=0 [ -n "$http_config_file" ] && [ -z "$(echo ${http_config_file} | grep $TMP_PATH)" ] && http_config_file=$TMP_PATH/$http_config_file if [ -n "$log_file" ] && [ -z "$(echo ${log_file} | grep $TMP_PATH)" ]; then log_file=$TMP_PATH/$log_file else log_file="/dev/null" fi local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z') local remarks=$(config_n_get $node remarks) local server_host=$(config_n_get $node address) local server_port=$(config_n_get $node port) [ -n "$relay_port" ] && { server_host="127.0.0.1" server_port=$relay_port } local error_msg tmp if [ -n "$server_host" ] && [ -n "$server_port" ]; then check_host $server_host [ $? != 0 ] && { echolog " - Socks节点:[$remarks]${server_host} 是非法的服务器地址,无法启动!" return 1 } tmp="${server_host}:${server_port}" else error_msg="某种原因,此 Socks 服务的相关配置已失联,启动中止!" fi if [ "$type" == "sing-box" ] || [ "$type" == "xray" ]; then local protocol=$(config_n_get $node protocol) if [ "$protocol" == "_balancing" ] || [ "$protocol" == "_shunt" ] || [ "$protocol" == "_iface" ] || [ "$protocol" == "_urltest" ]; then unset error_msg fi fi [ -n "${error_msg}" ] && { [ "$bind" != "127.0.0.1" ] && echolog " - Socks节点:[$remarks]${tmp},启动中止 ${bind}:${socks_port} ${error_msg}" return 1 } [ "$bind" != "127.0.0.1" ] && echolog " - Socks节点:[$remarks]${tmp},启动 ${bind}:${socks_port}" case "$type" in socks) local _socks_address=$(config_n_get $node address) local _socks_port=$(config_n_get $node port) local _socks_username=$(config_n_get $node username) local _socks_password=$(config_n_get $node password) [ "$http_port" != "0" ] && { http_flag=1 config_file="${config_file//SOCKS/HTTP_SOCKS}" local _extra_param="-local_http_address $bind -local_http_port $http_port" } local bin=$(first_type $(config_t_get global_app sing_box_file) sing-box) if [ -n "$bin" ]; then type="sing-box" lua $UTIL_SINGBOX gen_proto_config -local_socks_address $bind -local_socks_port $socks_port ${_extra_param} -server_proto socks -server_address ${_socks_address} -server_port ${_socks_port} -server_username ${_socks_username} -server_password ${_socks_password} > $config_file ln_run "$bin" ${type} $log_file run -c "$config_file" else bin=$(first_type $(config_t_get global_app xray_file) xray) [ -n "$bin" ] && { type="xray" lua $UTIL_XRAY gen_proto_config -local_socks_address $bind -local_socks_port $socks_port ${_extra_param} -server_proto socks -server_address ${_socks_address} -server_port ${_socks_port} -server_username ${_socks_username} -server_password ${_socks_password} > $config_file ln_run "$bin" ${type} $log_file run -c "$config_file" } fi ;; sing-box) [ "$http_port" != "0" ] && { http_flag=1 config_file="${config_file//SOCKS/HTTP_SOCKS}" local _args="http_address=$bind http_port=$http_port" } [ -n "$relay_port" ] && _args="${_args} server_host=$server_host server_port=$server_port" [ -n "$no_run" ] && _args="${_args} no_run=1" run_singbox flag=$flag node=$node socks_address=$bind socks_port=$socks_port config_file=$config_file log_file=$log_file ${_args} ;; xray) [ "$http_port" != "0" ] && { http_flag=1 config_file="${config_file//SOCKS/HTTP_SOCKS}" local _args="http_address=$bind http_port=$http_port" } [ -n "$relay_port" ] && _args="${_args} server_host=$server_host server_port=$server_port" [ -n "$no_run" ] && _args="${_args} no_run=1" run_xray flag=$flag node=$node socks_address=$bind socks_port=$socks_port config_file=$config_file log_file=$log_file ${_args} ;; trojan*) lua $UTIL_TROJAN gen_config -node $node -run_type client -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file [ -n "$no_run" ] || ln_run "$(first_type ${type})" "${type}" $log_file -c "$config_file" ;; naiveproxy) lua $UTIL_NAIVE gen_config -node $node -run_type socks -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file [ -n "$no_run" ] || ln_run "$(first_type naive)" naive $log_file "$config_file" ;; ssr) lua $UTIL_SS gen_config -node $node -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file [ -n "$no_run" ] || ln_run "$(first_type ssr-local)" "ssr-local" $log_file -c "$config_file" -v -u ;; ss) [ -n "$no_run" ] || { local plugin_sh="${config_file%.json}_plugin.sh" local _extra_param="-plugin_sh $plugin_sh" } lua $UTIL_SS gen_config -node $node -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port -mode tcp_and_udp ${_extra_param} > $config_file [ -n "$no_run" ] || ln_run "$(first_type ss-local)" "ss-local" $log_file -c "$config_file" -v ;; ss-rust) local _extra_param [ "$http_port" != "0" ] && { http_flag=1 config_file="${config_file//SOCKS/HTTP_SOCKS}" _extra_param="-local_http_address $bind -local_http_port $http_port" } [ -n "$no_run" ] || { local plugin_sh="${config_file%.json}_plugin.sh" _extra_param="${_extra_param:+$_extra_param }-plugin_sh $plugin_sh" } lua $UTIL_SS gen_config -node $node -local_socks_address $bind -local_socks_port $socks_port -server_host $server_host -server_port $server_port ${_extra_param} > $config_file [ -n "$no_run" ] || ln_run "$(first_type sslocal)" "sslocal" $log_file -c "$config_file" -v ;; hysteria2) [ "$http_port" != "0" ] && { http_flag=1 config_file="${config_file//SOCKS/HTTP_SOCKS}" local _extra_param="-local_http_address $bind -local_http_port $http_port" } lua $UTIL_HYSTERIA2 gen_config -node $node -local_socks_address $bind -local_socks_port $socks_port -server_host $server_host -server_port $server_port ${_extra_param} > $config_file [ -n "$no_run" ] || ln_run "$(first_type $(config_t_get global_app hysteria_file))" "hysteria" $log_file -c "$config_file" client ;; tuic) lua $UTIL_TUIC gen_config -node $node -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file [ -n "$no_run" ] || ln_run "$(first_type tuic-client)" "tuic-client" $log_file -c "$config_file" ;; esac set_cache_var "node_${node}_socks_port" "${socks_port}" # http to socks [ -z "$http_flag" ] && [ "$http_port" != "0" ] && [ -n "$http_config_file" ] && [ "$type" != "sing-box" ] && [ "$type" != "xray" ] && [ "$type" != "socks" ] && { local bin=$(first_type $(config_t_get global_app sing_box_file) sing-box) if [ -n "$bin" ]; then type="sing-box" lua $UTIL_SINGBOX gen_proto_config -local_http_address $bind -local_http_port $http_port -server_proto socks -server_address "127.0.0.1" -server_port $socks_port -server_username $_username -server_password $_password > $http_config_file [ -n "$no_run" ] || ln_run "$bin" ${type} /dev/null run -c "$http_config_file" else bin=$(first_type $(config_t_get global_app xray_file) xray) [ -n "$bin" ] && type="xray" [ -z "$type" ] && return 1 lua $UTIL_XRAY gen_proto_config local_http_address $bind -local_http_port $http_port -server_proto socks -server_address "127.0.0.1" -server_port $socks_port -server_username $_username -server_password $_password > $http_config_file [ -n "$no_run" ] || ln_run "$bin" ${type} /dev/null run -c "$http_config_file" fi } unset http_flag [ -z "$no_run" ] && [ "${server_host}" != "127.0.0.1" ] && [ "$type" != "sing-box" ] && [ "$type" != "xray" ] && echo "${node}" >> $TMP_PATH/direct_node_list } run_redir() { local node proto bind local_port config_file log_file eval_set_val $@ local tcp_node_socks_flag tcp_node_http_flag [ -n "$config_file" ] && [ -z "$(echo ${config_file} | grep $TMP_PATH)" ] && config_file=${GLOBAL_ACL_PATH}/${config_file} if [ -n "$log_file" ] && [ -z "$(echo ${log_file} | grep $TMP_PATH)" ]; then log_file=${GLOBAL_ACL_PATH}/${log_file} else log_file="/dev/null" fi local proto=$(echo $proto | tr 'A-Z' 'a-z') local PROTO=$(echo $proto | tr 'a-z' 'A-Z') local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z') local enable_log=$(config_t_get global log_${proto} 1) [ "$enable_log" != "1" ] && log_file="/dev/null" local remarks=$(config_n_get $node remarks) local server_host=$(config_n_get $node address) local port=$(config_n_get $node port) [ -n "$server_host" ] && [ -n "$port" ] && { check_host $server_host [ $? != 0 ] && { echolog "${PROTO}节点:[$remarks]${server_host} 是非法的服务器地址,无法启动!" return 1 } } [ "$bind" != "127.0.0.1" ] && echolog "${PROTO}节点:[$remarks],监听端口:$local_port" case "$PROTO" in UDP) case "$type" in socks) local _socks_address=$(config_n_get $node address) local _socks_port=$(config_n_get $node port) local _socks_username=$(config_n_get $node username) local _socks_password=$(config_n_get $node password) run_ipt2socks flag=default proto=UDP local_port=${local_port} socks_address=${_socks_address} socks_port=${_socks_port} socks_username=${_socks_username} socks_password=${_socks_password} log_file=${log_file} ;; sing-box) local protocol=$(config_n_get $node protocol) [ "$protocol" = "_shunt" ] && { local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat" local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat" if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then echolog "* 缺少Geo规则文件,UDP Sing-Box分流节点无法正常使用!" fi } run_singbox flag=UDP node=$node udp_redir_port=$local_port config_file=$config_file log_file=$log_file ;; xray) local protocol=$(config_n_get $node protocol) [ "$protocol" = "_shunt" ] && { local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat" local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat" if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then echolog "* 缺少Geo规则文件,UDP Xray分流节点无法正常使用!" fi } run_xray flag=UDP node=$node udp_redir_port=$local_port config_file=$config_file log_file=$log_file ;; trojan*) local loglevel=$(config_t_get global trojan_loglevel "2") lua $UTIL_TROJAN gen_config -node $node -run_type nat -local_addr "0.0.0.0" -local_port $local_port -loglevel $loglevel > $config_file ln_run "$(first_type ${type})" "${type}" $log_file -c "$config_file" ;; naiveproxy) echolog "Naiveproxy不支持UDP转发!" ;; ssr) lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port > $config_file ln_run "$(first_type ssr-redir)" "ssr-redir" $log_file -c "$config_file" -v -U ;; ss) local plugin_sh="${config_file%.json}_plugin.sh" local _extra_param="-plugin_sh $plugin_sh" lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port -mode udp_only $_extra_param > $config_file ln_run "$(first_type ss-redir)" "ss-redir" $log_file -c "$config_file" -v ;; ss-rust) local plugin_sh="${config_file%.json}_plugin.sh" local _extra_param="-plugin_sh $plugin_sh" lua $UTIL_SS gen_config -node $node -local_udp_redir_port $local_port $_extra_param > $config_file ln_run "$(first_type sslocal)" "sslocal" $log_file -c "$config_file" -v ;; hysteria2) lua $UTIL_HYSTERIA2 gen_config -node $node -local_udp_redir_port $local_port > $config_file ln_run "$(first_type $(config_t_get global_app hysteria_file))" "hysteria" $log_file -c "$config_file" client ;; tuic) echolog "TUIC不支持UDP转发!" ;; esac ;; TCP) [ "$TCP_UDP" = "1" ] && echolog "UDP节点:与TCP节点相同" tcp_node_socks=1 tcp_node_socks_bind_local=$(config_t_get global tcp_node_socks_bind_local 1) tcp_node_socks_bind="127.0.0.1" [ "${tcp_node_socks_bind_local}" != "1" ] && tcp_node_socks_bind="0.0.0.0" tcp_node_socks_port=$(get_new_port $(config_t_get global tcp_node_socks_port 1070)) tcp_node_http_port=$(config_t_get global tcp_node_http_port 0) [ "$tcp_node_http_port" != "0" ] && tcp_node_http=1 if [ $PROXY_IPV6 == "1" ]; then echolog "开启实验性IPv6透明代理(TProxy),请确认您的节点及类型支持IPv6!" fi if [ "${TCP_PROXY_WAY}" = "redirect" ]; then can_ipt=$(echo "$REDIRECT_LIST" | grep "$type") elif [ "${TCP_PROXY_WAY}" = "tproxy" ]; then can_ipt=$(echo "$TPROXY_LIST" | grep "$type") fi [ -z "$can_ipt" ] && type="socks" case "$type" in socks) _socks_flag=1 _socks_address=$(config_n_get $node address) _socks_port=$(config_n_get $node port) _socks_username=$(config_n_get $node username) _socks_password=$(config_n_get $node password) [ -z "$can_ipt" ] && { local _config_file=$config_file _config_file="TCP_SOCKS_${node}.json" local _port=$(get_new_port 2080) run_socks flag="TCP" node=$node bind=127.0.0.1 socks_port=${_port} config_file=${_config_file} _socks_address=127.0.0.1 _socks_port=${_port} unset _socks_username unset _socks_password } ;; sing-box) local _flag="TCP" local _args="" [ "$tcp_node_socks" = "1" ] && { tcp_node_socks_flag=1 _args="${_args} socks_address=${tcp_node_socks_bind} socks_port=${tcp_node_socks_port}" config_file="${config_file//TCP/TCP_SOCKS}" } [ "$tcp_node_http" = "1" ] && { tcp_node_http_flag=1 _args="${_args} http_port=${tcp_node_http_port}" config_file="${config_file//TCP/TCP_HTTP}" } [ "$TCP_UDP" = "1" ] && { UDP_REDIR_PORT=$local_port unset UDP_NODE _flag="TCP_UDP" _args="${_args} udp_redir_port=${UDP_REDIR_PORT}" config_file="${config_file//TCP/TCP_UDP}" } local protocol=$(config_n_get $node protocol) local default_node=$(config_n_get $node default_node) local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp) [ "${DNS_MODE}" != "sing-box" ] && [ "${DNS_MODE}" != "udp" ] && [ "$protocol" = "_shunt" ] && [ "$default_node" = "_direct" ] && { DNS_MODE="sing-box" v2ray_dns_mode="tcp" echolog "* 当前TCP节点采用Sing-Box分流且默认节点为直连,远程DNS过滤模式将默认使用Sing-Box(TCP),防止环回!" } [ "$protocol" = "_shunt" ] && { local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat" local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat" if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then echolog "* 缺少Geo规则文件,TCP Sing-Box分流节点无法正常使用!" fi } [ "${DNS_MODE}" = "sing-box" ] && { NO_PLUGIN_DNS=1 config_file="${config_file//.json/_DNS.json}" _args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}" DNSMASQ_FILTER_PROXY_IPV6=0 local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip) [ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}" [ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0" resolve_dns_port=${NEXT_DNS_LISTEN_PORT} _args="${_args} dns_listen_port=${resolve_dns_port}" case "$(config_t_get global direct_dns_mode "auto")" in udp) _args="${_args} direct_dns_udp_server=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g')" ;; tcp) _args="${_args} direct_dns_tcp_server=$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g')" ;; esac _args="${_args} remote_dns_protocol=${v2ray_dns_mode}" case "$v2ray_dns_mode" in tcp) _args="${_args} remote_dns_tcp_server=${REMOTE_DNS}" resolve_dns_log="Sing-Box DNS(127.0.0.1#${resolve_dns_port}) -> tcp://${REMOTE_DNS}" ;; doh) remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query") _args="${_args} remote_dns_doh=${remote_dns_doh}" resolve_dns_log="Sing-Box DNS(127.0.0.1#${resolve_dns_port}) -> ${remote_dns_doh}" ;; esac local remote_fakedns=$(config_t_get global remote_fakedns 0) [ "${remote_fakedns}" = "1" ] && { fakedns=1 _args="${_args} remote_fakedns=1" resolve_dns_log="${resolve_dns_log} + FakeDNS" } NEXT_DNS_LISTEN_PORT=$(expr $NEXT_DNS_LISTEN_PORT + 1) } run_singbox flag=$_flag node=$node tcp_redir_port=$local_port tcp_proxy_way=$TCP_PROXY_WAY config_file=$config_file log_file=$log_file ${_args} ;; xray) local _flag="TCP" local _args="" [ "$tcp_node_socks" = "1" ] && { tcp_node_socks_flag=1 _args="${_args} socks_address=${tcp_node_socks_bind} socks_port=${tcp_node_socks_port}" config_file="${config_file//TCP/TCP_SOCKS}" } [ "$tcp_node_http" = "1" ] && { tcp_node_http_flag=1 _args="${_args} http_port=${tcp_node_http_port}" config_file="${config_file//TCP/TCP_HTTP}" } [ "$TCP_UDP" = "1" ] && { UDP_REDIR_PORT=$local_port unset UDP_NODE _flag="TCP_UDP" _args="${_args} udp_redir_port=${UDP_REDIR_PORT}" config_file="${config_file//TCP/TCP_UDP}" } local protocol=$(config_n_get $node protocol) local default_node=$(config_n_get $node default_node) local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp) [ "${DNS_MODE}" != "xray" ] && [ "${DNS_MODE}" != "udp" ] && [ "$protocol" = "_shunt" ] && [ "$default_node" = "_direct" ] && { DNS_MODE="xray" v2ray_dns_mode="tcp" echolog "* 当前TCP节点采用Xray分流且默认节点为直连,远程DNS过滤模式将默认使用Xray(TCP),防止环回!" } [ "$protocol" = "_shunt" ] && { local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat" local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat" if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then echolog "* 缺少Geo规则文件,TCP Xray分流节点无法正常使用!" fi } [ "${DNS_MODE}" = "xray" ] && { NO_PLUGIN_DNS=1 config_file="${config_file//.json/_DNS.json}" _args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}" DNSMASQ_FILTER_PROXY_IPV6=0 local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip) [ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}" [ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0" resolve_dns_port=${NEXT_DNS_LISTEN_PORT} _args="${_args} dns_listen_port=${resolve_dns_port}" case "$(config_t_get global direct_dns_mode "auto")" in udp) _args="${_args} direct_dns_udp_server=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g')" ;; tcp) _args="${_args} direct_dns_tcp_server=$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g')" ;; esac _args="${_args} remote_dns_tcp_server=${REMOTE_DNS}" if [ "$v2ray_dns_mode" = "tcp+doh" ]; then remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query") _args="${_args} remote_dns_doh=${remote_dns_doh}" resolve_dns_log="Xray DNS(127.0.0.1#${resolve_dns_port}) -> (${remote_dns_doh})(A/AAAA) + tcp://${REMOTE_DNS}" else resolve_dns_log="Xray DNS(127.0.0.1#${resolve_dns_port}) -> tcp://${REMOTE_DNS}" fi local remote_fakedns=$(config_t_get global remote_fakedns 0) [ "${remote_fakedns}" = "1" ] && { fakedns=1 _args="${_args} remote_fakedns=1" resolve_dns_log="${resolve_dns_log} + FakeDNS" } NEXT_DNS_LISTEN_PORT=$(expr $NEXT_DNS_LISTEN_PORT + 1) } run_xray flag=$_flag node=$node tcp_redir_port=$local_port tcp_proxy_way=$TCP_PROXY_WAY config_file=$config_file log_file=$log_file ${_args} ;; trojan*) [ "${TCP_PROXY_WAY}" = "tproxy" ] && lua_tproxy_arg="-use_tproxy true" [ "$TCP_UDP" = "1" ] && { config_file="${config_file//TCP/TCP_UDP}" UDP_REDIR_PORT=$TCP_REDIR_PORT unset UDP_NODE } local loglevel=$(config_t_get global trojan_loglevel "2") lua $UTIL_TROJAN gen_config -node $node -run_type nat -local_addr "0.0.0.0" -local_port $local_port -loglevel $loglevel $lua_tproxy_arg > $config_file ln_run "$(first_type ${type})" "${type}" $log_file -c "$config_file" ;; naiveproxy) lua $UTIL_NAIVE gen_config -node $node -run_type redir -local_addr "0.0.0.0" -local_port $local_port > $config_file ln_run "$(first_type naive)" naive $log_file "$config_file" ;; ssr) [ "${TCP_PROXY_WAY}" = "tproxy" ] && lua_tproxy_arg="-tcp_tproxy true" [ "$TCP_UDP" = "1" ] && { config_file="${config_file//TCP/TCP_UDP}" UDP_REDIR_PORT=$TCP_REDIR_PORT unset UDP_NODE _extra_param="-u" } lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port $lua_tproxy_arg > $config_file ln_run "$(first_type ssr-redir)" "ssr-redir" $log_file -c "$config_file" -v ${_extra_param} ;; ss) [ "${TCP_PROXY_WAY}" = "tproxy" ] && lua_tproxy_arg="-tcp_tproxy true" local lua_mode_arg="-mode tcp_only" [ "$TCP_UDP" = "1" ] && { config_file="${config_file//TCP/TCP_UDP}" UDP_REDIR_PORT=$TCP_REDIR_PORT unset UDP_NODE lua_mode_arg="-mode tcp_and_udp" } local plugin_sh="${config_file%.json}_plugin.sh" lua_mode_arg="${lua_mode_arg} -plugin_sh $plugin_sh" lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port $lua_mode_arg $lua_tproxy_arg > $config_file ln_run "$(first_type ss-redir)" "ss-redir" $log_file -c "$config_file" -v ;; ss-rust) local _extra_param="-local_tcp_redir_port $local_port" [ "${TCP_PROXY_WAY}" = "tproxy" ] && _extra_param="${_extra_param} -tcp_tproxy true" [ "$tcp_node_socks" = "1" ] && { tcp_node_socks_flag=1 config_file="${config_file//TCP/TCP_SOCKS}" _extra_param="${_extra_param} -local_socks_address ${tcp_node_socks_bind} -local_socks_port ${tcp_node_socks_port}" } [ "$tcp_node_http" = "1" ] && { tcp_node_http_flag=1 config_file="${config_file//TCP/TCP_HTTP}" _extra_param="${_extra_param} -local_http_port ${tcp_node_http_port}" } [ "$TCP_UDP" = "1" ] && { config_file="${config_file//TCP/TCP_UDP}" UDP_REDIR_PORT=$TCP_REDIR_PORT unset UDP_NODE _extra_param="${_extra_param} -local_udp_redir_port $local_port" } local plugin_sh="${config_file%.json}_plugin.sh" _extra_param="${_extra_param} -plugin_sh $plugin_sh" lua $UTIL_SS gen_config -node $node ${_extra_param} > $config_file ln_run "$(first_type sslocal)" "sslocal" $log_file -c "$config_file" -v ;; hysteria2) local _extra_param="-local_tcp_redir_port $local_port" [ "$tcp_node_socks" = "1" ] && { tcp_node_socks_flag=1 config_file="${config_file//TCP/TCP_SOCKS}" _extra_param="${_extra_param} -local_socks_address ${tcp_node_socks_bind} -local_socks_port ${tcp_node_socks_port}" } [ "$tcp_node_http" = "1" ] && { tcp_node_http_flag=1 config_file="${config_file//TCP/TCP_HTTP}" _extra_param="${_extra_param} -local_http_port ${tcp_node_http_port}" } [ "$TCP_UDP" = "1" ] && { config_file="${config_file//TCP/TCP_UDP}" UDP_REDIR_PORT=$TCP_REDIR_PORT unset UDP_NODE _extra_param="${_extra_param} -local_udp_redir_port $local_port" } _extra_param="${_extra_param} -tcp_proxy_way ${TCP_PROXY_WAY}" lua $UTIL_HYSTERIA2 gen_config -node $node ${_extra_param} > $config_file ln_run "$(first_type $(config_t_get global_app hysteria_file))" "hysteria" $log_file -c "$config_file" client ;; esac if [ -n "${_socks_flag}" ]; then local _flag="TCP" [ "$TCP_UDP" = "1" ] && { _flag="TCP_UDP" UDP_REDIR_PORT=$TCP_REDIR_PORT unset UDP_NODE } local _socks_tproxy="" [ "${TCP_PROXY_WAY}" = "tproxy" ] && _socks_tproxy="1" run_ipt2socks flag=default proto=${_flag} tcp_tproxy=${_socks_tproxy} local_port=${local_port} socks_address=${_socks_address} socks_port=${_socks_port} socks_username=${_socks_username} socks_password=${_socks_password} log_file=${log_file} fi [ -z "$tcp_node_socks_flag" ] && { [ "$tcp_node_socks" = "1" ] && { local config_file="SOCKS_TCP.json" local log_file="SOCKS_TCP.log" local http_port=0 local http_config_file="HTTP2SOCKS_TCP.json" [ "$tcp_node_http" = "1" ] && [ -z "$tcp_node_http_flag" ] && { http_port=$tcp_node_http_port } run_socks flag=TCP node=$node bind=$tcp_node_socks_bind socks_port=$tcp_node_socks_port config_file=$config_file http_port=$http_port http_config_file=$http_config_file } } [ "$tcp_node_socks" = "1" ] && { TCP_SOCKS_server="127.0.0.1:$tcp_node_socks_port" set_cache_var "GLOBAL_TCP_SOCKS_server" "${TCP_SOCKS_server}" } ;; esac unset tcp_node_socks_flag tcp_node_http_flag [ "$type" != "sing-box" ] && [ "$type" != "xray" ] && echo "${node}" >> $TMP_PATH/direct_node_list [ -n "${redir_port}" ] && set_cache_var "node_${node}_${PROTO}_redir_port" "${local_port}" } start_redir() { local proto=${1} eval node=\$${proto}_NODE if [ -n "$node" ]; then TYPE=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z') local config_file="${proto}.json" local log_file="${proto}.log" eval current_port=\$${proto}_REDIR_PORT local port=$(echo $(get_new_port $current_port $proto)) eval ${proto}_REDIR_PORT=$port run_redir node=$node proto=${proto} bind=0.0.0.0 local_port=$port config_file=$config_file log_file=$log_file set_cache_var "ACL_GLOBAL_${proto}_node" "${node}" set_cache_var "ACL_GLOBAL_${proto}_redir_port" "${port}" [ "$TCP_UDP" = "1" ] && { set_cache_var "ACL_GLOBAL_UDP_node" "${node}" set_cache_var "ACL_GLOBAL_UDP_redir_port" "${port}" } else [ "${proto}" = "UDP" ] && [ "$TCP_UDP" = "1" ] && return echolog "${proto}节点没有选择或为空,不代理${proto}。" fi } start_socks() { [ "$SOCKS_ENABLED" = "1" ] && { local ids=$(uci show $CONFIG | grep "=socks" | awk -F '.' '{print $2}' | awk -F '=' '{print $1}') [ -n "$ids" ] && { echolog "分析 Socks 服务的节点配置..." for id in $ids; do local enabled=$(config_n_get $id enabled 0) [ "$enabled" == "0" ] && continue local node=$(config_n_get $id node) [ -z "$node" ] && continue local bind_local=$(config_n_get $id bind_local 0) local bind="0.0.0.0" [ "$bind_local" = "1" ] && bind="127.0.0.1" local port=$(config_n_get $id port) local config_file="SOCKS_${id}.json" local log_file="SOCKS_${id}.log" local log=$(config_n_get $id log 1) [ "$log" == "0" ] && log_file="" local http_port=$(config_n_get $id http_port 0) local http_config_file="HTTP2SOCKS_${id}.json" run_socks flag=$id node=$node bind=$bind socks_port=$port config_file=$config_file http_port=$http_port http_config_file=$http_config_file log_file=$log_file set_cache_var "socks_${id}" "$node" #自动切换逻辑 local enable_autoswitch=$(config_n_get $id enable_autoswitch 0) [ "$enable_autoswitch" = "1" ] && $APP_PATH/socks_auto_switch.sh ${id} > /dev/null 2>&1 & done } } } socks_node_switch() { local flag new_node eval_set_val $@ [ -n "$flag" ] && [ -n "$new_node" ] && { local prefix pf filename # 结束 SS 插件进程 for prefix in "" "HTTP_"; do pf="$TMP_PATH/${prefix}SOCKS_${flag}_plugin.pid" [ -s "$pf" ] && kill -9 "$(head -n1 "$pf")" >/dev/null 2>&1 done pgrep -af "$TMP_BIN_PATH" | awk -v P1="${flag}" 'BEGIN{IGNORECASE=1}$0~P1 && !/acl\/|acl_/{print $1}' | xargs kill -9 >/dev/null 2>&1 for prefix in "" "HTTP_" "HTTP2"; do rm -rf "$TMP_PATH/${prefix}SOCKS_${flag}"* done for filename in $(ls ${TMP_SCRIPT_FUNC_PATH}); do cmd=$(cat ${TMP_SCRIPT_FUNC_PATH}/${filename}) [ -n "$(echo $cmd | grep "${flag}")" ] && rm -f ${TMP_SCRIPT_FUNC_PATH}/${filename} done local bind_local=$(config_n_get $flag bind_local 0) local bind="0.0.0.0" [ "$bind_local" = "1" ] && bind="127.0.0.1" local port=$(config_n_get $flag port) local config_file="SOCKS_${flag}.json" local log_file="SOCKS_${flag}.log" local log=$(config_n_get $flag log 1) [ "$log" == "0" ] && log_file="" local http_port=$(config_n_get $flag http_port 0) local http_config_file="HTTP2SOCKS_${flag}.json" LOG_FILE="/dev/null" run_socks flag=$flag node=$new_node bind=$bind socks_port=$port config_file=$config_file http_port=$http_port http_config_file=$http_config_file log_file=$log_file set_cache_var "socks_${flag}" "$new_node" local USE_TABLES=$(get_cache_var "USE_TABLES") [ -n "$USE_TABLES" ] && source $APP_PATH/${USE_TABLES}.sh filter_direct_node_list } } clean_log() { logsnum=$(cat $LOG_FILE 2>/dev/null | wc -l) [ "$logsnum" -gt 1000 ] && { echo "" > $LOG_FILE echolog "日志文件过长,清空处理!" } } clean_crontab() { [ -f "/tmp/lock/${CONFIG}_cron.lock" ] && return touch /etc/crontabs/root #sed -i "/${CONFIG}/d" /etc/crontabs/root >/dev/null 2>&1 sed -i "/$(echo "/etc/init.d/${CONFIG}" | sed 's#\/#\\\/#g')/d" /etc/crontabs/root >/dev/null 2>&1 sed -i "/$(echo "lua ${APP_PATH}/rule_update.lua log" | sed 's#\/#\\\/#g')/d" /etc/crontabs/root >/dev/null 2>&1 sed -i "/$(echo "lua ${APP_PATH}/subscribe.lua start" | sed 's#\/#\\\/#g')/d" /etc/crontabs/root >/dev/null 2>&1 pgrep -af "${CONFIG}/" | awk '/tasks\.sh/{print $1}' | xargs kill -9 >/dev/null 2>&1 rm -rf /tmp/lock/${CONFIG}_tasks.lock } start_crontab() { if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then start_daemon=$(config_t_get global_delay start_daemon 0) [ "$start_daemon" = "1" ] && $APP_PATH/monitor.sh > /dev/null 2>&1 & fi [ -f "/tmp/lock/${CONFIG}_cron.lock" ] && { rm -rf "/tmp/lock/${CONFIG}_cron.lock" echolog "当前为计划任务自动运行,不重新配置定时任务。" return } clean_crontab [ "$ENABLED" != 1 ] && { /etc/init.d/cron restart return } stop_week_mode=$(config_t_get global_delay stop_week_mode) stop_time_mode=$(config_t_get global_delay stop_time_mode) if [ -n "$stop_week_mode" ]; then local t="0 $stop_time_mode * * $stop_week_mode" [ "$stop_week_mode" = "7" ] && t="0 $stop_time_mode * * *" if [ "$stop_week_mode" = "8" ]; then update_loop=1 else echo "$t /etc/init.d/$CONFIG stop > /dev/null 2>&1 &" >>/etc/crontabs/root fi echolog "配置定时任务:自动关闭服务。" fi start_week_mode=$(config_t_get global_delay start_week_mode) start_time_mode=$(config_t_get global_delay start_time_mode) if [ -n "$start_week_mode" ]; then local t="0 $start_time_mode * * $start_week_mode" [ "$start_week_mode" = "7" ] && t="0 $start_time_mode * * *" if [ "$start_week_mode" = "8" ]; then update_loop=1 else echo "$t /etc/init.d/$CONFIG start > /dev/null 2>&1 &" >>/etc/crontabs/root fi echolog "配置定时任务:自动开启服务。" fi restart_week_mode=$(config_t_get global_delay restart_week_mode) restart_time_mode=$(config_t_get global_delay restart_time_mode) if [ -n "$restart_week_mode" ]; then local t="0 $restart_time_mode * * $restart_week_mode" [ "$restart_week_mode" = "7" ] && t="0 $restart_time_mode * * *" if [ "$restart_week_mode" = "8" ]; then update_loop=1 else echo "$t /etc/init.d/$CONFIG restart > /dev/null 2>&1 &" >>/etc/crontabs/root fi echolog "配置定时任务:自动重启服务。" fi autoupdate=$(config_t_get global_rules auto_update) weekupdate=$(config_t_get global_rules week_update) dayupdate=$(config_t_get global_rules time_update) if [ "$autoupdate" = "1" ]; then local t="0 $dayupdate * * $weekupdate" [ "$weekupdate" = "7" ] && t="0 $dayupdate * * *" if [ "$weekupdate" = "8" ]; then update_loop=1 else echo "$t lua $APP_PATH/rule_update.lua log all cron > /dev/null 2>&1 &" >>/etc/crontabs/root fi echolog "配置定时任务:自动更新规则。" fi TMP_SUB_PATH=$TMP_PATH/sub_crontabs mkdir -p $TMP_SUB_PATH for item in $(uci show ${CONFIG} | grep "=subscribe_list" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do if [ "$(config_n_get $item auto_update 0)" = "1" ]; then cfgid=$(uci show ${CONFIG}.$item | head -n 1 | cut -d '.' -sf 2 | cut -d '=' -sf 1) remark=$(config_n_get $item remark) week_update=$(config_n_get $item week_update) time_update=$(config_n_get $item time_update) echo "$cfgid" >> $TMP_SUB_PATH/${week_update}_${time_update} echolog "配置定时任务:自动更新【$remark】订阅。" fi done [ -d "${TMP_SUB_PATH}" ] && { for name in $(ls ${TMP_SUB_PATH}); do week_update=$(echo $name | awk -F '_' '{print $1}') time_update=$(echo $name | awk -F '_' '{print $2}') cfgids=$(echo -n $(cat ${TMP_SUB_PATH}/${name}) | sed 's# #,#g') local t="0 $time_update * * $week_update" [ "$week_update" = "7" ] && t="0 $time_update * * *" if [ "$week_update" = "8" ]; then update_loop=1 else echo "$t lua $APP_PATH/subscribe.lua start $cfgids cron > /dev/null 2>&1 &" >>/etc/crontabs/root fi done rm -rf $TMP_SUB_PATH } if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then [ "$update_loop" = "1" ] && { $APP_PATH/tasks.sh > /dev/null 2>&1 & echolog "自动更新:启动循环更新进程。" } else echolog "运行于非代理模式,仅允许服务启停的定时任务。" fi /etc/init.d/cron restart } stop_crontab() { [ -f "/tmp/lock/${CONFIG}_cron.lock" ] && return clean_crontab /etc/init.d/cron restart #echolog "清除定时执行命令。" } start_dns() { echolog "DNS域名解析:" local china_ng_local_dns=$(IFS=','; set -- $LOCAL_DNS; [ "${1%%[#:]*}" = "127.0.0.1" ] && echo "$1" || ([ -n "$2" ] && echo "$1,$2" || echo "$1")) local sing_box_local_dns= local direct_dns_mode=$(config_t_get global direct_dns_mode "auto") #获取访问控制节点所使用的DNS分流模式 local ACL_RULE_DNSMASQ=0 for acl_section in $(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do if [ "$(config_n_get $acl_section enabled)" = "1" ] && \ [ "$(config_n_get $acl_section dns_shunt)" = "dnsmasq" ]; then ACL_RULE_DNSMASQ=1 break fi done case "$direct_dns_mode" in udp) LOCAL_DNS=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g') china_ng_local_dns=${LOCAL_DNS} sing_box_local_dns="direct_dns_udp_server=${LOCAL_DNS}" ;; tcp) local DIRECT_DNS=$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g') china_ng_local_dns="tcp://${DIRECT_DNS}" sing_box_local_dns="direct_dns_tcp_server=${DIRECT_DNS}" #当全局(包括访问控制节点)开启chinadns-ng时,不启动新进程。 [ "$DNS_SHUNT" != "chinadns-ng" ] || [ "$ACL_RULE_DNSMASQ" = "1" ] && { LOCAL_DNS="127.0.0.1#${NEXT_DNS_LISTEN_PORT}" local china_ng_c_dns="tcp://$(get_first_dns DIRECT_DNS 53 | sed 's/:/#/g')" ln_run "$(first_type chinadns-ng)" chinadns-ng "/dev/null" -b :: -l ${NEXT_DNS_LISTEN_PORT} -c ${china_ng_c_dns} -d chn echolog " - ChinaDNS-NG(${LOCAL_DNS}) -> ${china_ng_c_dns}" echolog " * 请确保上游直连 DNS 支持 TCP 查询。" NEXT_DNS_LISTEN_PORT=$(expr $NEXT_DNS_LISTEN_PORT + 1) } ;; auto) #Automatic logic is already done by default : ;; esac # 追加直连DNS到iptables/nftables [ "$(config_t_get global_haproxy balancing_enable 0)" != "1" ] && IPT_APPEND_DNS= add_default_port() { [ -z "$1" ] && echo "" || echo "$1" | awk -F',' '{for(i=1;i<=NF;i++){if($i !~ /#/) $i=$i"#53";} print $0;}' OFS=',' } LOCAL_DNS=$(add_default_port "$LOCAL_DNS") IPT_APPEND_DNS=$(add_default_port "${IPT_APPEND_DNS:-$LOCAL_DNS}") echo "$IPT_APPEND_DNS" | grep -q -E "(^|,)$LOCAL_DNS(,|$)" || IPT_APPEND_DNS="${IPT_APPEND_DNS:+$IPT_APPEND_DNS,}$LOCAL_DNS" [ -n "$DIRECT_DNS" ] && { DIRECT_DNS=$(add_default_port "$DIRECT_DNS") echo "$IPT_APPEND_DNS" | grep -q -E "(^|,)$DIRECT_DNS(,|$)" || IPT_APPEND_DNS="${IPT_APPEND_DNS:+$IPT_APPEND_DNS,}$DIRECT_DNS" } # 排除127.0.0.1的条目 IPT_APPEND_DNS=$(echo "$IPT_APPEND_DNS" | awk -F',' '{for(i=1;i<=NF;i++) if($i !~ /^127\.0\.0\.1/) printf (i>1?",":"") $i; print ""}' | sed 's/^,\|,$//g') TUN_DNS="127.0.0.1#${NEXT_DNS_LISTEN_PORT}" [ -n "${NO_PLUGIN_DNS}" ] && TUN_DNS="127.0.0.1#${resolve_dns_port}" case "$DNS_MODE" in dns2socks) local dns2socks_socks_server=$(echo $(config_t_get global socks_server 127.0.0.1:1080) | sed "s/#/:/g") local dns2socks_forward=$(get_first_dns REMOTE_DNS 53 | sed 's/#/:/g') run_dns2socks socks=$dns2socks_socks_server listen_address=127.0.0.1 listen_port=${NEXT_DNS_LISTEN_PORT} dns=$dns2socks_forward cache=$DNS_CACHE echolog " - dns2socks(${TUN_DNS}),${dns2socks_socks_server} -> tcp://${dns2socks_forward}" ;; sing-box) [ -z "${NO_PLUGIN_DNS}" ] && { local config_file=$TMP_PATH/DNS.json local log_file=$TMP_PATH/DNS.log local log_file=/dev/null local _args="type=$DNS_MODE config_file=$config_file log_file=$log_file" [ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0" _args="${_args} direct_dns_query_strategy=${DIRECT_DNS_QUERY_STRATEGY}" _args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}" DNSMASQ_FILTER_PROXY_IPV6=0 local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip) [ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}" TCP_PROXY_DNS=1 local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp) _args="${_args} dns_listen_port=${NEXT_DNS_LISTEN_PORT}" _args="${_args} remote_dns_protocol=${v2ray_dns_mode}" case "$v2ray_dns_mode" in tcp) _args="${_args} remote_dns_tcp_server=${REMOTE_DNS}" echolog " - Sing-Box DNS(${TUN_DNS}) -> tcp://${REMOTE_DNS}" ;; doh) remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query") _args="${_args} remote_dns_doh=${remote_dns_doh}" echolog " - Sing-Box DNS(${TUN_DNS}) -> ${remote_dns_doh}" local _doh_url _doh_host _doh_port _doh_bootstrap parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap [ -n "${_doh_bootstrap}" ] && REMOTE_DNS="${_doh_bootstrap}#${_doh_port}" ;; esac _args="${_args} dns_socks_address=127.0.0.1 dns_socks_port=${tcp_node_socks_port}" [ -n "${sing_box_local_dns}" ] && _args="${_args} ${sing_box_local_dns}" run_singbox ${_args} } ;; xray) [ -z "${NO_PLUGIN_DNS}" ] && { local config_file=$TMP_PATH/DNS.json local log_file=$TMP_PATH/DNS.log local log_file=/dev/null local _args="type=$DNS_MODE config_file=$config_file log_file=$log_file" [ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0" _args="${_args} direct_dns_query_strategy=${DIRECT_DNS_QUERY_STRATEGY}" _args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}" DNSMASQ_FILTER_PROXY_IPV6=0 local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip) [ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}" TCP_PROXY_DNS=1 _args="${_args} dns_listen_port=${NEXT_DNS_LISTEN_PORT}" _args="${_args} remote_dns_tcp_server=${REMOTE_DNS}" local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp) if [ "$v2ray_dns_mode" = "tcp+doh" ]; then remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query") _args="${_args} remote_dns_doh=${remote_dns_doh}" echolog " - Xray DNS(${TUN_DNS}) -> (${remote_dns_doh})(A/AAAA) + tcp://${REMOTE_DNS}" local _doh_url _doh_host _doh_port _doh_bootstrap parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap [ -n "${_doh_bootstrap}" ] && REMOTE_DNS="${REMOTE_DNS},${_doh_bootstrap}#${_doh_port}" else echolog " - Xray DNS(${TUN_DNS}) -> tcp://${REMOTE_DNS}" fi _args="${_args} dns_socks_address=127.0.0.1 dns_socks_port=${tcp_node_socks_port}" run_xray ${_args} } ;; udp) UDP_PROXY_DNS=1 local china_ng_listen_port=${NEXT_DNS_LISTEN_PORT} local china_ng_trust_dns="udp://$(get_first_dns REMOTE_DNS 53 | sed 's/:/#/g')" if [ "$DNS_SHUNT" != "chinadns-ng" ] && [ "$FILTER_PROXY_IPV6" = "1" ]; then DNSMASQ_FILTER_PROXY_IPV6=0 local no_ipv6_trust="-N" ln_run "$(first_type chinadns-ng)" chinadns-ng "/dev/null" -b :: -l ${china_ng_listen_port} -t ${china_ng_trust_dns} -d gfw ${no_ipv6_trust} echolog " - ChinaDNS-NG(${TUN_DNS}) -> ${china_ng_trust_dns}" else TUN_DNS="$(echo ${REMOTE_DNS} | sed 's/#/:/g' | sed -E 's/\:([^:]+)$/#\1/g')" echolog " - udp://${TUN_DNS}" fi ;; tcp) TCP_PROXY_DNS=1 local china_ng_listen_port=${NEXT_DNS_LISTEN_PORT} local china_ng_trust_dns="tcp://$(get_first_dns REMOTE_DNS 53 | sed 's/:/#/g')" [ "$DNS_SHUNT" != "chinadns-ng" ] && { [ "$FILTER_PROXY_IPV6" = "1" ] && DNSMASQ_FILTER_PROXY_IPV6=0 && local no_ipv6_trust="-N" ln_run "$(first_type chinadns-ng)" chinadns-ng "/dev/null" -b :: -l ${china_ng_listen_port} -t ${china_ng_trust_dns} -d gfw ${no_ipv6_trust} echolog " - ChinaDNS-NG(${TUN_DNS}) -> ${china_ng_trust_dns}" } ;; esac [ -n "${resolve_dns_log}" ] && echolog " - ${resolve_dns_log}" [ -n "${TCP_PROXY_DNS}" ] && echolog " * 请确认上游 DNS 支持 TCP/DoH 查询,如非直连地址,确保 TCP 代理打开,并且已经正确转发!" [ -n "${UDP_PROXY_DNS}" ] && echolog " * 请确认上游 DNS 支持 UDP 查询并已使用 UDP 节点,如上游 DNS 非直连地址,确保 UDP 代理打开,并且已经正确转发!" [ "${DNS_SHUNT}" = "smartdns" ] && { if command -v smartdns > /dev/null 2>&1; then rm -rf $TMP_PATH2/dnsmasq_default* local group_domestic=$(config_t_get global group_domestic) local smartdns_remote_dns=$(config_t_get global smartdns_remote_dns) if [ -n "${smartdns_remote_dns}" -a "${smartdns_remote_dns}" != "nil" ]; then smartdns_remote_dns=$(echo ${smartdns_remote_dns} | tr -s ' ' '|') else smartdns_remote_dns="tcp://1.1.1.1" fi local subnet_ip=$(config_t_get global remote_dns_client_ip) lua $APP_PATH/helper_smartdns_add.lua -FLAG "default" -SMARTDNS_CONF "/tmp/etc/smartdns/$CONFIG.conf" \ -LOCAL_GROUP ${group_domestic:-nil} -REMOTE_GROUP "passwall_proxy" -REMOTE_PROXY_SERVER ${TCP_SOCKS_server} -USE_DEFAULT_DNS "${USE_DEFAULT_DNS:-direct}" \ -REMOTE_DNS ${smartdns_remote_dns} -DNS_MODE ${DNS_MODE:-socks} -TUN_DNS ${TUN_DNS} -REMOTE_FAKEDNS ${fakedns:-0} \ -USE_DIRECT_LIST "${USE_DIRECT_LIST}" -USE_PROXY_LIST "${USE_PROXY_LIST}" -USE_BLOCK_LIST "${USE_BLOCK_LIST}" -USE_GFW_LIST "${USE_GFW_LIST}" -CHN_LIST "${CHN_LIST}" \ -TCP_NODE ${TCP_NODE} -DEFAULT_PROXY_MODE "${TCP_PROXY_MODE}" -NO_PROXY_IPV6 ${FILTER_PROXY_IPV6:-0} -NFTFLAG ${nftflag:-0} \ -SUBNET ${subnet_ip:-0} -NO_LOGIC_LOG ${NO_LOGIC_LOG:-0} source $APP_PATH/helper_smartdns.sh restart echolog " - 域名解析:使用SmartDNS,请确保配置正常。" return else DNS_SHUNT="dnsmasq" echolog " * 未安装SmartDNS,默认使用Dnsmasq进行域名解析!" fi } [ "$DNS_SHUNT" = "chinadns-ng" ] && [ -n "$(first_type chinadns-ng)" ] && { chinadns_ng_min=2024.04.13 chinadns_ng_now=$($(first_type chinadns-ng) -V | grep -i "ChinaDNS-NG " | awk '{print $2}') if [ $(check_ver "$chinadns_ng_now" "$chinadns_ng_min") = 1 ]; then echolog " * 注意:当前 ChinaDNS-NG 版本为[ $chinadns_ng_now ],请更新到[ $chinadns_ng_min ]或以上版本,否则 DNS 有可能无法正常工作!" fi [ "$FILTER_PROXY_IPV6" = "1" ] && DNSMASQ_FILTER_PROXY_IPV6=0 [ -z "${china_ng_listen_port}" ] && local china_ng_listen_port=$(expr $NEXT_DNS_LISTEN_PORT + 1) local china_ng_listen="127.0.0.1#${china_ng_listen_port}" [ -z "${china_ng_trust_dns}" ] && local china_ng_trust_dns=${TUN_DNS} echolog " - ChinaDNS-NG(${china_ng_listen}):直连DNS:${china_ng_local_dns},可信DNS:${china_ng_trust_dns}" china_ng_listen="${china_ng_listen},::1#${china_ng_listen_port}" run_chinadns_ng \ _flag="default" \ _listen_port=${china_ng_listen_port} \ _dns_local=${china_ng_local_dns} \ _dns_trust=${china_ng_trust_dns} \ _no_ipv6_trust=${FILTER_PROXY_IPV6} \ _use_direct_list=${USE_DIRECT_LIST} \ _use_proxy_list=${USE_PROXY_LIST} \ _use_block_list=${USE_BLOCK_LIST} \ _gfwlist=${USE_GFW_LIST} \ _chnlist=${CHN_LIST} \ _default_mode=${TCP_PROXY_MODE} \ _default_tag=$(config_t_get global chinadns_ng_default_tag smart) \ _no_logic_log=0 \ _tcp_node=${TCP_NODE} \ _remote_fakedns=${fakedns:-0} USE_DEFAULT_DNS="chinadns_ng" } [ "$USE_DEFAULT_DNS" = "remote" ] && { dnsmasq_version=$(dnsmasq -v | grep -i "Dnsmasq version " | awk '{print $3}') [ "$(expr $dnsmasq_version \>= 2.87)" == 0 ] && echolog "Dnsmasq版本低于2.87,有可能无法正常使用!!!" } local DNSMASQ_TUN_DNS=$(get_first_dns TUN_DNS 53 | sed 's/:/#/g') local RUN_NEW_DNSMASQ=1 RUN_NEW_DNSMASQ=${DNS_REDIRECT} if [ "${RUN_NEW_DNSMASQ}" == "0" ]; then #The old logic will be removed in the future. #Run a copy dnsmasq instance, DNS hijack that don't need a proxy devices. [ "1" = "0" ] && { DIRECT_DNSMASQ_PORT=$(get_new_port 11400) DIRECT_DNSMASQ_CONF=${GLOBAL_ACL_PATH}/direct_dnsmasq.conf DIRECT_DNSMASQ_CONF_PATH=${GLOBAL_ACL_PATH}/direct_dnsmasq.d mkdir -p ${DIRECT_DNSMASQ_CONF_PATH} lua $APP_PATH/helper_dnsmasq.lua copy_instance -LISTEN_PORT ${DIRECT_DNSMASQ_PORT} -DNSMASQ_CONF ${DIRECT_DNSMASQ_CONF} -TMP_DNSMASQ_PATH ${DIRECT_DNSMASQ_CONF_PATH} ln_run "$(first_type dnsmasq)" "dnsmasq_direct" "/dev/null" -C ${DIRECT_DNSMASQ_CONF} -x ${GLOBAL_ACL_PATH}/direct_dnsmasq.pid echo "${DIRECT_DNSMASQ_PORT}" > ${GLOBAL_ACL_PATH}/direct_dnsmasq_port } #Rewrite the default DNS service configuration #Modify the default dnsmasq service lua $APP_PATH/helper_dnsmasq.lua stretch lua $APP_PATH/helper_dnsmasq.lua add_rule -FLAG "default" -TMP_DNSMASQ_PATH ${GLOBAL_DNSMASQ_CONF_PATH} -DNSMASQ_CONF_FILE ${GLOBAL_DNSMASQ_CONF} \ -DEFAULT_DNS ${DEFAULT_DNS} -LOCAL_DNS ${LOCAL_DNS} -TUN_DNS ${DNSMASQ_TUN_DNS} \ -REMOTE_FAKEDNS ${fakedns:-0} -USE_DEFAULT_DNS "${USE_DEFAULT_DNS:-direct}" -CHINADNS_DNS ${china_ng_listen:-0} \ -USE_DIRECT_LIST "${USE_DIRECT_LIST}" -USE_PROXY_LIST "${USE_PROXY_LIST}" -USE_BLOCK_LIST "${USE_BLOCK_LIST}" -USE_GFW_LIST "${USE_GFW_LIST}" -CHN_LIST "${CHN_LIST}" \ -TCP_NODE ${TCP_NODE} -DEFAULT_PROXY_MODE ${TCP_PROXY_MODE} -NO_PROXY_IPV6 ${DNSMASQ_FILTER_PROXY_IPV6:-0} -NFTFLAG ${nftflag:-0} \ -NO_LOGIC_LOG ${NO_LOGIC_LOG:-0} uci -q add_list dhcp.@dnsmasq[0].addnmount=${GLOBAL_DNSMASQ_CONF_PATH} uci -q commit dhcp lua $APP_PATH/helper_dnsmasq.lua logic_restart -LOG 1 else #Run a copy dnsmasq instance, DNS hijack for that need proxy devices. GLOBAL_DNSMASQ_PORT=$(get_new_port 11400) GLOBAL_DNSMASQ_CONF=${GLOBAL_ACL_PATH}/dnsmasq.conf GLOBAL_DNSMASQ_CONF_PATH=${GLOBAL_ACL_PATH}/dnsmasq.d lua $APP_PATH/helper_dnsmasq.lua add_rule -FLAG "default" -TMP_DNSMASQ_PATH ${GLOBAL_DNSMASQ_CONF_PATH} -DNSMASQ_CONF_FILE ${GLOBAL_DNSMASQ_CONF} \ -LISTEN_PORT ${GLOBAL_DNSMASQ_PORT} -DEFAULT_DNS ${DEFAULT_DNS} -LOCAL_DNS ${LOCAL_DNS} -TUN_DNS ${DNSMASQ_TUN_DNS} \ -REMOTE_FAKEDNS ${fakedns:-0} -USE_DEFAULT_DNS "${USE_DEFAULT_DNS:-direct}" -CHINADNS_DNS ${china_ng_listen:-0} \ -USE_DIRECT_LIST "${USE_DIRECT_LIST}" -USE_PROXY_LIST "${USE_PROXY_LIST}" -USE_BLOCK_LIST "${USE_BLOCK_LIST}" -USE_GFW_LIST "${USE_GFW_LIST}" -CHN_LIST "${CHN_LIST}" \ -TCP_NODE ${TCP_NODE} -DEFAULT_PROXY_MODE ${TCP_PROXY_MODE} -NO_PROXY_IPV6 ${DNSMASQ_FILTER_PROXY_IPV6:-0} -NFTFLAG ${nftflag:-0} \ -NO_LOGIC_LOG ${NO_LOGIC_LOG:-0} ln_run "$(first_type dnsmasq)" "dnsmasq_default" "/dev/null" -C ${GLOBAL_DNSMASQ_CONF} -x ${GLOBAL_ACL_PATH}/dnsmasq.pid set_cache_var "ACL_default_dns_port" "${GLOBAL_DNSMASQ_PORT}" DNS_REDIRECT_PORT=${GLOBAL_DNSMASQ_PORT} #dhcp.leases to hosts $APP_PATH/lease2hosts.sh > /dev/null 2>&1 & fi } add_ip2route() { local ip=$(get_host_ip "ipv4" $1) [ -z "$ip" ] && { echolog " - 无法解析[${1}],路由表添加失败!" return 1 } local remarks="${1}" [ "$remarks" != "$ip" ] && remarks="${1}(${ip})" . /lib/functions/network.sh local gateway device network_get_gateway gateway "$2" network_get_device device "$2" [ -z "${device}" ] && device="$2" if [ -n "${gateway}" ]; then route add -host ${ip} gw ${gateway} dev ${device} >/dev/null 2>&1 echo "$ip" >> $TMP_ROUTE_PATH/${device} echolog " - [${remarks}]添加到接口[${device}]路由表成功!" else echolog " - [${remarks}]添加到接口[${device}]路由表失功!原因是找不到[${device}]网关。" fi } delete_ip2route() { [ -d "${TMP_ROUTE_PATH}" ] && { local interface for interface in $(ls ${TMP_ROUTE_PATH}); do for ip in $(cat ${TMP_ROUTE_PATH}/${interface}); do route del -host ${ip} dev ${interface} >/dev/null 2>&1 done done } } start_haproxy() { [ "$(config_t_get global_haproxy balancing_enable 0)" != "1" ] && return haproxy_path=$TMP_PATH/haproxy haproxy_conf="config.cfg" lua $APP_PATH/haproxy.lua -path ${haproxy_path} -conf ${haproxy_conf} -dns ${LOCAL_DNS} ln_run "$(first_type haproxy)" haproxy "/dev/null" -f "${haproxy_path}/${haproxy_conf}" } kill_all() { kill -9 $(pidof "$@") >/dev/null 2>&1 } acl_app() { local items=$(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1) [ -n "$items" ] && { local item local socks_port redir_port dns_port dnsmasq_port chinadns_port local msg msg2 socks_port=11100 redir_port=11200 dns_port=11300 dnsmasq_port=${GLOBAL_DNSMASQ_PORT:-11400} chinadns_port=11500 for item in $items; do local sid=$(uci -q show "${CONFIG}.${item}" | grep "=acl_rule" | awk -F '=' '{print $1}' | awk -F '.' '{print $2}') [ "$(config_n_get $sid enabled)" = "1" ] || continue eval $(uci -q show "${CONFIG}.${item}" | cut -d'.' -sf 3-) if [ -n "${sources}" ]; then for s in $sources; do local s2 is_iprange=$(lua_api "iprange(\"${s}\")") if [ "${is_iprange}" = "true" ]; then s2="iprange:${s}" elif [ -n "$(echo ${s} | grep '^ipset:')" ]; then s2="ipset:${s}" else _ip_or_mac=$(lua_api "ip_or_mac(\"${s}\")") if [ "${_ip_or_mac}" = "ip" ]; then s2="ip:${s}" elif [ "${_ip_or_mac}" = "mac" ]; then s2="mac:${s}" fi fi [ -n "${s2}" ] && source_list="${source_list}\n${s2}" unset s2 done else source_list="any" fi local acl_path=${TMP_ACL_PATH}/$sid mkdir -p ${acl_path} [ -n "${source_list}" ] && echo -e "${source_list}" | sed '/^$/d' > ${acl_path}/source_list use_global_config=${use_global_config:-0} [ "${use_global_config}" = "1" ] && { tcp_node="default" udp_node="default" } tcp_no_redir_ports=${tcp_no_redir_ports:-${TCP_NO_REDIR_PORTS}} udp_no_redir_ports=${udp_no_redir_ports:-${UDP_NO_REDIR_PORTS}} if has_1_65535 "$tcp_no_redir_ports" && has_1_65535 "$udp_no_redir_ports"; then unset use_global_config unset tcp_node unset udp_node else use_direct_list=${use_direct_list:-1} use_proxy_list=${use_proxy_list:-1} use_block_list=${use_block_list:-1} use_gfw_list=${use_gfw_list:-1} chn_list=${chn_list:-direct} tcp_proxy_mode=${tcp_proxy_mode:-proxy} udp_proxy_mode=${udp_proxy_mode:-proxy} filter_proxy_ipv6=${filter_proxy_ipv6:-0} dnsmasq_filter_proxy_ipv6=${filter_proxy_ipv6} dns_shunt=${dns_shunt:-dnsmasq} dns_mode=${dns_mode:-dns2socks} remote_dns=${remote_dns:-1.1.1.1} use_default_dns=${use_default_dns:-direct} [ "$dns_mode" = "sing-box" ] && { [ "$v2ray_dns_mode" = "doh" ] && remote_dns=${remote_dns_doh:-https://1.1.1.1/dns-query} } fi [ -n "$tcp_node" ] && { local GLOBAL_TCP_NODE=$(get_cache_var "ACL_GLOBAL_TCP_node") [ -n "${GLOBAL_TCP_NODE}" ] && GLOBAL_TCP_redir_port=$(get_cache_var "ACL_GLOBAL_TCP_redir_port") if [ "$tcp_node" = "default" ]; then if [ -n "${GLOBAL_TCP_NODE}" ]; then set_cache_var "ACL_${sid}_tcp_node" "${GLOBAL_TCP_NODE}" set_cache_var "ACL_${sid}_tcp_redir_port" "${GLOBAL_TCP_redir_port}" set_cache_var "ACL_${sid}_dns_port" "${GLOBAL_DNSMASQ_PORT}" set_cache_var "ACL_${sid}_tcp_default" "1" else echolog " - 全局节点未启用,跳过【${remarks}】" fi else [ "$(config_get_type $tcp_node)" = "nodes" ] && { if [ -n "${GLOBAL_TCP_NODE}" ] && [ "$tcp_node" = "${GLOBAL_TCP_NODE}" ]; then set_cache_var "ACL_${sid}_tcp_node" "${GLOBAL_TCP_NODE}" set_cache_var "ACL_${sid}_tcp_redir_port" "${GLOBAL_TCP_redir_port}" set_cache_var "ACL_${sid}_dns_port" "${GLOBAL_DNSMASQ_PORT}" set_cache_var "ACL_${sid}_tcp_default" "1" else run_dns() { local _dns_port [ -n $1 ] && _dns_port=$1 [ -z ${_dns_port} ] && { dns_port=$(get_new_port $(expr $dns_port + 1)) _dns_port=$dns_port if [ "$dns_mode" = "dns2socks" ]; then run_dns2socks flag=acl_${sid} socks_address=127.0.0.1 socks_port=$socks_port listen_address=0.0.0.0 listen_port=${_dns_port} dns=$remote_dns cache=1 elif [ "$dns_mode" = "sing-box" -o "$dns_mode" = "xray" ]; then config_file=$TMP_ACL_PATH/${tcp_node}_SOCKS_${socks_port}_DNS.json [ "$dns_mode" = "xray" ] && [ "$v2ray_dns_mode" = "tcp+doh" ] && remote_dns_doh=${remote_dns_doh:-https://1.1.1.1/dns-query} local type=${dns_mode} [ "${dns_mode}" = "sing-box" ] && type="singbox" dnsmasq_filter_proxy_ipv6=0 remote_dns_query_strategy="UseIP" [ "$filter_proxy_ipv6" = "1" ] && remote_dns_query_strategy="UseIPv4" run_${type} flag=acl_${sid} type=$dns_mode dns_socks_address=127.0.0.1 dns_socks_port=$socks_port dns_listen_port=${_dns_port} remote_dns_protocol=${v2ray_dns_mode} remote_dns_tcp_server=${remote_dns} remote_dns_doh="${remote_dns_doh}" remote_dns_query_strategy=${remote_dns_query_strategy} remote_dns_client_ip=${remote_dns_client_ip} config_file=$config_file fi set_cache_var "node_${tcp_node}_$(echo -n "${remote_dns}" | md5sum | cut -d " " -f1)" "${_dns_port}" } [ "$dns_shunt" = "chinadns-ng" ] && [ -n "$(first_type chinadns-ng)" ] && { chinadns_ng_min=2024.04.13 chinadns_ng_now=$($(first_type chinadns-ng) -V | grep -i "ChinaDNS-NG " | awk '{print $2}') if [ $(check_ver "$chinadns_ng_now" "$chinadns_ng_min") = 1 ]; then echolog " * 注意:当前 ChinaDNS-NG 版本为[ $chinadns_ng_now ],请更新到[ $chinadns_ng_min ]或以上版本,否则 DNS 有可能无法正常工作!" fi [ "$filter_proxy_ipv6" = "1" ] && dnsmasq_filter_proxy_ipv6=0 chinadns_port=$(expr $chinadns_port + 1) _china_ng_listen="127.0.0.1#${chinadns_port},::1#${chinadns_port}" _chinadns_local_dns=$(IFS=','; set -- $LOCAL_DNS; [ "${1%%[#:]*}" = "127.0.0.1" ] && echo "$1" || ([ -n "$2" ] && echo "$1,$2" || echo "$1")) _direct_dns_mode=$(config_t_get global direct_dns_mode "auto") case "${_direct_dns_mode}" in udp) _chinadns_local_dns=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g') ;; tcp) _chinadns_local_dns="tcp://$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g')" ;; esac run_chinadns_ng \ _flag="$sid" \ _listen_port=${chinadns_port} \ _dns_local=${_chinadns_local_dns} \ _dns_trust=127.0.0.1#${_dns_port} \ _no_ipv6_trust=${filter_proxy_ipv6} \ _use_direct_list=${use_direct_list} \ _use_proxy_list=${use_proxy_list} \ _use_block_list=${use_block_list} \ _gfwlist=${use_gfw_list} \ _chnlist=${chn_list} \ _default_mode=${tcp_proxy_mode} \ _default_tag=${chinadns_ng_default_tag:-smart} \ _no_logic_log=1 \ _tcp_node=${tcp_node} \ _remote_fakedns=0 use_default_dns="chinadns_ng" } dnsmasq_port=$(get_new_port $(expr $dnsmasq_port + 1)) local dnsmasq_conf=${acl_path}/dnsmasq.conf local dnsmasq_conf_path=${acl_path}/dnsmasq.d lua $APP_PATH/helper_dnsmasq.lua add_rule -FLAG ${sid} -TMP_DNSMASQ_PATH ${dnsmasq_conf_path} -DNSMASQ_CONF_FILE ${dnsmasq_conf} \ -LISTEN_PORT ${dnsmasq_port} -DEFAULT_DNS ${DEFAULT_DNS} -LOCAL_DNS $LOCAL_DNS \ -USE_DIRECT_LIST "${use_direct_list}" -USE_PROXY_LIST "${use_proxy_list}" -USE_BLOCK_LIST "${use_block_list}" -USE_GFW_LIST "${use_gfw_list}" -CHN_LIST "${chn_list}" \ -TUN_DNS "127.0.0.1#${_dns_port}" -REMOTE_FAKEDNS 0 -USE_DEFAULT_DNS "${use_default_dns:-direct}" -CHINADNS_DNS ${_china_ng_listen:-0} \ -TCP_NODE $tcp_node -DEFAULT_PROXY_MODE ${tcp_proxy_mode} -NO_PROXY_IPV6 ${dnsmasq_filter_proxy_ipv6:-0} -NFTFLAG ${nftflag:-0} \ -NO_LOGIC_LOG 1 ln_run "$(first_type dnsmasq)" "dnsmasq_${sid}" "/dev/null" -C ${dnsmasq_conf} -x ${acl_path}/dnsmasq.pid set_cache_var "ACL_${sid}_dns_port" "${dnsmasq_port}" set_cache_var "node_${tcp_node}_$(echo -n "${tcp_proxy_mode}${remote_dns}" | md5sum | cut -d " " -f1)" "${dnsmasq_port}" #dhcp.leases to hosts $APP_PATH/lease2hosts.sh > /dev/null 2>&1 & } _redir_port=$(get_cache_var "node_${tcp_node}_redir_port") _socks_port=$(get_cache_var "node_${tcp_node}_socks_port") if [ -n "${_socks_port}" ] && [ -n "${_redir_port}" ]; then socks_port=${_socks_port} tcp_port=${_redir_port} _dnsmasq_port=$(get_cache_var "node_${tcp_node}_$(echo -n "${tcp_proxy_mode}${remote_dns}" | md5sum | cut -d " " -f1)") if [ -z "${_dnsmasq_port}" ]; then _dns_port=$(get_cache_var "node_${tcp_node}_$(echo -n "${remote_dns}" | md5sum | cut -d " " -f1)") run_dns ${_dns_port} else [ -n "${_dnsmasq_port}" ] && set_cache_var "ACL_${sid}_dns_port" "${_dnsmasq_port}" fi else socks_port=$(get_new_port $(expr $socks_port + 1)) set_cache_var "node_${tcp_node}_socks_port" "${socks_port}" redir_port=$(get_new_port $(expr $redir_port + 1)) set_cache_var "node_${tcp_node}_redir_port" "${redir_port}" tcp_port=$redir_port local type=$(echo $(config_n_get $tcp_node type) | tr 'A-Z' 'a-z') if [ -n "${type}" ] && ([ "${type}" = "sing-box" ] || [ "${type}" = "xray" ]); then config_file="acl/${tcp_node}_TCP_${redir_port}.json" _extra_param="socks_address=127.0.0.1 socks_port=$socks_port" _extra_param="${_extra_param} tcp_proxy_way=$TCP_PROXY_WAY" if [ "$dns_mode" = "sing-box" ] || [ "$dns_mode" = "xray" ]; then dns_port=$(get_new_port $(expr $dns_port + 1)) _dns_port=$dns_port config_file="${config_file//TCP_/DNS_${_dns_port}_TCP_}" remote_dns_doh=${remote_dns} dnsmasq_filter_proxy_ipv6=0 remote_dns_query_strategy="UseIP" [ "$filter_proxy_ipv6" = "1" ] && remote_dns_query_strategy="UseIPv4" [ "$dns_mode" = "xray" ] && [ "$v2ray_dns_mode" = "tcp+doh" ] && remote_dns_doh=${remote_dns_doh:-https://1.1.1.1/dns-query} _extra_param="dns_listen_port=${_dns_port} remote_dns_protocol=${v2ray_dns_mode} remote_dns_tcp_server=${remote_dns} remote_dns_doh=${remote_dns_doh} remote_dns_query_strategy=${remote_dns_query_strategy} remote_dns_client_ip=${remote_dns_client_ip}" fi [ -n "$udp_node" ] && ([ "$udp_node" = "tcp" ] || [ "$udp_node" = "$tcp_node" ]) && { config_file="${config_file//TCP_/TCP_UDP_}" _extra_param="${_extra_param} udp_redir_port=$redir_port" } config_file="$TMP_PATH/$config_file" [ "${type}" = "sing-box" ] && type="singbox" run_${type} flag=$tcp_node node=$tcp_node tcp_redir_port=$redir_port ${_extra_param} config_file=$config_file else config_file="acl/${tcp_node}_SOCKS_${socks_port}.json" run_socks flag=$tcp_node node=$tcp_node bind=127.0.0.1 socks_port=$socks_port config_file=$config_file local log_file=$TMP_ACL_PATH/ipt2socks_${tcp_node}_${redir_port}.log log_file="/dev/null" run_ipt2socks flag=acl_${tcp_node} tcp_tproxy=${is_tproxy} local_port=$redir_port socks_address=127.0.0.1 socks_port=$socks_port log_file=$log_file fi run_dns ${_dns_port} fi set_cache_var "ACL_${sid}_tcp_node" "${tcp_node}" set_cache_var "ACL_${sid}_tcp_redir_port" "${tcp_port}" fi } fi } [ -n "$udp_node" ] && { if [ "$udp_node" = "default" ]; then local GLOBAL_UDP_NODE=$(get_cache_var "ACL_GLOBAL_UDP_node") [ -n "${GLOBAL_UDP_NODE}" ] && GLOBAL_UDP_redir_port=$(get_cache_var "ACL_GLOBAL_UDP_redir_port") if [ -n "${GLOBAL_UDP_NODE}" ]; then set_cache_var "ACL_${sid}_udp_node" "${GLOBAL_UDP_NODE}" set_cache_var "ACL_${sid}_udp_redir_port" "${GLOBAL_UDP_redir_port}" set_cache_var "ACL_${sid}_udp_default" "1" else echolog " - 全局节点未启用,跳过【${remarks}】" fi elif [ "$udp_node" = "tcp" ] || [ "$udp_node" = "$tcp_node" ]; then udp_node=$(get_cache_var "ACL_${sid}_tcp_node") udp_port=$(get_cache_var "ACL_${sid}_tcp_redir_port") set_cache_var "ACL_${sid}_udp_node" "${udp_node}" set_cache_var "ACL_${sid}_udp_redir_port" "${udp_port}" else [ "$(config_get_type $udp_node)" = "nodes" ] && { if [ -n "${GLOBAL_UDP_NODE}" ] && [ "$udp_node" = "${GLOBAL_UDP_NODE}" ]; then set_cache_var "ACL_${sid}_udp_node" "${GLOBAL_UDP_NODE}" set_cache_var "ACL_${sid}_udp_redir_port" "${GLOBAL_UDP_redir_port}" set_cache_var "ACL_${sid}_udp_default" "1" else _redir_port=$(get_cache_var "node_${udp_node}_redir_port") _socks_port=$(get_cache_var "node_${udp_node}_socks_port") if [ -n "${_socks_port}" ] && [ -n "${_redir_port}" ]; then socks_port=${_socks_port} udp_port=${_redir_port} else socks_port=$(get_new_port $(expr $socks_port + 1)) set_cache_var "node_${udp_node}_socks_port" "${socks_port}" redir_port=$(get_new_port $(expr $redir_port + 1)) set_cache_var "node_${udp_node}_redir_port" "${redir_port}" udp_port=$redir_port local type=$(echo $(config_n_get $udp_node type) | tr 'A-Z' 'a-z') if [ -n "${type}" ] && ([ "${type}" = "sing-box" ] || [ "${type}" = "xray" ]); then config_file="acl/${udp_node}_UDP_${redir_port}.json" config_file="$TMP_PATH/$config_file" [ "${type}" = "sing-box" ] && type="singbox" run_${type} flag=$udp_node node=$udp_node udp_redir_port=$redir_port config_file=$config_file else config_file="acl/${udp_node}_SOCKS_${socks_port}.json" run_socks flag=$udp_node node=$udp_node bind=127.0.0.1 socks_port=$socks_port config_file=$config_file local log_file=$TMP_ACL_PATH/ipt2socks_${udp_node}_${redir_port}.log log_file="/dev/null" run_ipt2socks flag=acl_${udp_node} local_port=$redir_port socks_address=127.0.0.1 socks_port=$socks_port log_file=$log_file fi fi set_cache_var "ACL_${sid}_udp_node" "${udp_node}" set_cache_var "ACL_${sid}_udp_redir_port" "${udp_port}" fi } fi } unset enabled sid remarks sources interface tcp_no_redir_ports udp_no_redir_ports use_global_config tcp_node udp_node use_direct_list use_proxy_list use_block_list use_gfw_list chn_list tcp_proxy_mode udp_proxy_mode filter_proxy_ipv6 dns_mode remote_dns v2ray_dns_mode remote_dns_doh remote_dns_client_ip unset _ip _mac _iprange _ipset _ip_or_mac source_list tcp_port udp_port config_file _extra_param unset _china_ng_listen _chinadns_local_dns _direct_dns_mode chinadns_ng_default_tag dnsmasq_filter_proxy_ipv6 done unset socks_port redir_port dns_port dnsmasq_port chinadns_port } } start() { mkdir -p /tmp/etc /tmp/log $TMP_PATH $TMP_BIN_PATH $TMP_SCRIPT_FUNC_PATH $TMP_ROUTE_PATH $TMP_ACL_PATH $TMP_PATH2 get_config export V2RAY_LOCATION_ASSET=$(config_t_get global_rules v2ray_location_asset "/usr/share/v2ray/") export XRAY_LOCATION_ASSET=$V2RAY_LOCATION_ASSET export ENABLE_DEPRECATED_GEOSITE=true export ENABLE_DEPRECATED_GEOIP=true ulimit -n 65535 start_haproxy start_socks nftflag=0 local use_nft=$(config_t_get global_forwarding use_nft 0) local USE_TABLES if [ "$use_nft" == 0 ]; then if [ -n "$(command -v iptables-legacy || command -v iptables)" ] && [ -n "$(command -v ipset)" ] && [ -n "$(dnsmasq --version | grep 'Compile time options:.* ipset')" ]; then USE_TABLES="iptables" else echolog "系统未安装iptables或ipset或Dnsmasq没有开启ipset支持,无法使用iptables+ipset透明代理!" if [ -n "$(command -v fw4)" ] && [ -n "$(command -v nft)" ] && [ -n "$(dnsmasq --version | grep 'Compile time options:.* nftset')" ]; then echolog "检测到fw4,使用nftables进行透明代理。" USE_TABLES="nftables" nftflag=1 config_t_set global_forwarding use_nft 1 uci -q commit ${CONFIG} fi fi else if [ -n "$(dnsmasq --version | grep 'Compile time options:.* nftset')" ]; then USE_TABLES="nftables" nftflag=1 else echolog "Dnsmasq软件包不满足nftables透明代理要求,如需使用请确保dnsmasq版本在2.87以上并开启nftset支持。" fi fi check_depends $USE_TABLES [ "$USE_TABLES" = "nftables" ] && { dnsmasq_version=$(dnsmasq -v | grep -i "Dnsmasq version " | awk '{print $3}') [ "$(expr $dnsmasq_version \>= 2.90)" == 0 ] && echolog "Dnsmasq版本低于2.90,建议升级至2.90及以上版本以避免部分情况下Dnsmasq崩溃问题!" } if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then [ "$(uci -q get dhcp.@dnsmasq[0].dns_redirect)" == "1" ] && { uci -q set ${CONFIG}.@global[0].dnsmasq_dns_redirect='1' uci -q commit ${CONFIG} uci -q set dhcp.@dnsmasq[0].dns_redirect='0' uci -q commit dhcp lua $APP_PATH/helper_dnsmasq.lua restart -LOG 0 } fi [ "$ENABLED_DEFAULT_ACL" == 1 ] && { mkdir -p ${GLOBAL_ACL_PATH} start_redir TCP start_redir UDP start_dns } [ -n "$USE_TABLES" ] && source $APP_PATH/${USE_TABLES}.sh start set_cache_var "USE_TABLES" "$USE_TABLES" if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then bridge_nf_ipt=$(sysctl -e -n net.bridge.bridge-nf-call-iptables) set_cache_var "bak_bridge_nf_ipt" "$bridge_nf_ipt" sysctl -w net.bridge.bridge-nf-call-iptables=0 >/dev/null 2>&1 [ "$PROXY_IPV6" == "1" ] && { bridge_nf_ip6t=$(sysctl -e -n net.bridge.bridge-nf-call-ip6tables) set_cache_var "bak_bridge_nf_ip6t" "$bridge_nf_ip6t" sysctl -w net.bridge.bridge-nf-call-ip6tables=0 >/dev/null 2>&1 } fi start_crontab echolog "运行完成!\n" } stop() { clean_log eval_cache_var [ -n "$USE_TABLES" ] && source $APP_PATH/${USE_TABLES}.sh stop delete_ip2route # 结束 SS 插件进程 # kill_all xray-plugin v2ray-plugin obfs-local shadow-tls local pid_file pid find "$TMP_PATH" -type f -name '*_plugin.pid' | while read -r pid_file; do read -r pid < "$pid_file" if [ -n "$pid" ]; then kill -9 "$pid" >/dev/null 2>&1 fi done pgrep -f "sleep.*(6s|9s|58s)" | xargs kill -9 >/dev/null 2>&1 pgrep -af "${CONFIG}/" | awk '! /app\.sh|subscribe\.lua|rule_update\.lua|tasks\.sh|ujail/{print $1}' | xargs kill -9 >/dev/null 2>&1 stop_crontab source $APP_PATH/helper_smartdns.sh del rm -rf $GLOBAL_DNSMASQ_CONF rm -rf $GLOBAL_DNSMASQ_CONF_PATH [ "1" = "1" ] && { #restore logic bak_dnsmasq_dns_redirect=$(config_t_get global dnsmasq_dns_redirect) [ -n "${bak_dnsmasq_dns_redirect}" ] && { uci -q set dhcp.@dnsmasq[0].dns_redirect="${bak_dnsmasq_dns_redirect}" uci -q commit dhcp uci -q delete ${CONFIG}.@global[0].dnsmasq_dns_redirect uci -q commit ${CONFIG} } if [ -z "${ACL_default_dns_port}" ] || [ -n "${bak_dnsmasq_dns_redirect}" ]; then uci -q del_list dhcp.@dnsmasq[0].addnmount="${GLOBAL_DNSMASQ_CONF_PATH}" uci -q commit dhcp lua $APP_PATH/helper_dnsmasq.lua restart -LOG 0 fi [ -n "${bak_bridge_nf_ipt}" ] && sysctl -w net.bridge.bridge-nf-call-iptables=${bak_bridge_nf_ipt} >/dev/null 2>&1 [ -n "${bak_bridge_nf_ip6t}" ] && sysctl -w net.bridge.bridge-nf-call-ip6tables=${bak_bridge_nf_ip6t} >/dev/null 2>&1 } rm -rf $TMP_PATH rm -rf /tmp/lock/${CONFIG}_socks_auto_switch* rm -rf /tmp/lock/${CONFIG}_lease2hosts* echolog "清空并关闭相关程序和缓存完成。" exit 0 } get_config() { ENABLED_DEFAULT_ACL=0 TCP_REDIR_PORT=1041 UDP_REDIR_PORT=1051 ENABLED=$(config_t_get global enabled 0) SOCKS_ENABLED=$(config_t_get global socks_enabled 0) TCP_NODE=$(config_t_get global tcp_node) UDP_NODE=$(config_t_get global udp_node) TCP_UDP=0 if [ "$UDP_NODE" == "tcp" ]; then UDP_NODE=$TCP_NODE TCP_UDP=1 elif [ "$UDP_NODE" == "$TCP_NODE" ]; then TCP_UDP=1 fi [ "$ENABLED" == 1 ] && { [ -n "$TCP_NODE" ] && [ "$(config_get_type $TCP_NODE)" == "nodes" ] && ENABLED_DEFAULT_ACL=1 [ -n "$UDP_NODE" ] && [ "$(config_get_type $UDP_NODE)" == "nodes" ] && ENABLED_DEFAULT_ACL=1 } ENABLED_ACLS=$(config_t_get global acl_enable 0) [ "$ENABLED_ACLS" == 1 ] && { [ "$(uci show ${CONFIG} | grep "@acl_rule" | grep "enabled='1'" | wc -l)" == 0 ] && ENABLED_ACLS=0 } TCP_PROXY_WAY=$(config_t_get global_forwarding tcp_proxy_way redirect) PROXY_IPV6=$(config_t_get global_forwarding ipv6_tproxy 0) TCP_REDIR_PORTS=$(config_t_get global_forwarding tcp_redir_ports '80,443') UDP_REDIR_PORTS=$(config_t_get global_forwarding udp_redir_ports '1:65535') TCP_NO_REDIR_PORTS=$(config_t_get global_forwarding tcp_no_redir_ports 'disable') UDP_NO_REDIR_PORTS=$(config_t_get global_forwarding udp_no_redir_ports 'disable') TCP_PROXY_DROP_PORTS=$(config_t_get global_forwarding tcp_proxy_drop_ports 'disable') UDP_PROXY_DROP_PORTS=$(config_t_get global_forwarding udp_proxy_drop_ports '80,443') USE_DIRECT_LIST=$(config_t_get global use_direct_list 1) USE_PROXY_LIST=$(config_t_get global use_proxy_list 1) USE_BLOCK_LIST=$(config_t_get global use_block_list 1) USE_GFW_LIST=$(config_t_get global use_gfw_list 1) CHN_LIST=$(config_t_get global chn_list direct) TCP_PROXY_MODE=$(config_t_get global tcp_proxy_mode proxy) UDP_PROXY_MODE=$(config_t_get global udp_proxy_mode proxy) [ "${TCP_PROXY_MODE}" != "disable" ] && TCP_PROXY_MODE="proxy" [ "${UDP_PROXY_MODE}" != "disable" ] && UDP_PROXY_MODE="proxy" LOCALHOST_PROXY=$(config_t_get global localhost_proxy 1) [ "${LOCALHOST_PROXY}" == 1 ] && { LOCALHOST_TCP_PROXY_MODE=$TCP_PROXY_MODE LOCALHOST_UDP_PROXY_MODE=$UDP_PROXY_MODE } CLIENT_PROXY=$(config_t_get global client_proxy 1) DNS_SHUNT=$(config_t_get global dns_shunt dnsmasq) [ -z "$(first_type $DNS_SHUNT)" ] && DNS_SHUNT="dnsmasq" DNS_MODE=$(config_t_get global dns_mode tcp) [ "$DNS_SHUNT" = "smartdns" ] && DNS_MODE=$(config_t_get global smartdns_dns_mode socks) REMOTE_DNS=$(config_t_get global remote_dns 1.1.1.1:53 | sed 's/#/:/g' | sed -E 's/\:([^:]+)$/#\1/g') USE_DEFAULT_DNS=$(config_t_get global use_default_dns direct) FILTER_PROXY_IPV6=$(config_t_get global filter_proxy_ipv6 0) DNS_REDIRECT=$(config_t_get global dns_redirect 1) REDIRECT_LIST="socks ss ss-rust ssr sing-box xray trojan-plus naiveproxy hysteria2" TPROXY_LIST="socks ss ss-rust ssr sing-box xray trojan-plus hysteria2" NEXT_DNS_LISTEN_PORT=15353 TUN_DNS="127.0.0.1#${NEXT_DNS_LISTEN_PORT}" DNS_CACHE=0 DIRECT_DNS_QUERY_STRATEGY="UseIP" REMOTE_DNS_QUERY_STRATEGY="UseIP" [ "$FILTER_PROXY_IPV6" = "1" ] && REMOTE_DNS_QUERY_STRATEGY="UseIPv4" DNSMASQ_FILTER_PROXY_IPV6=${FILTER_PROXY_IPV6} RESOLVFILE=/tmp/resolv.conf.d/resolv.conf.auto [ -f "${RESOLVFILE}" ] && [ -s "${RESOLVFILE}" ] || RESOLVFILE=/tmp/resolv.conf.auto ISP_DNS=$(cat $RESOLVFILE 2>/dev/null | grep -E -o "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+" | grep -v -E '^(0\.0\.0\.0|127\.0\.0\.1)$' | awk '!seen[$0]++') ISP_DNS6=$(cat $RESOLVFILE 2>/dev/null | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | awk -F % '{print $1}' | awk -F " " '{print $2}' | grep -v -Fx ::1 | grep -v -Fx :: | awk '!seen[$0]++') DEFAULT_DNS=$(uci show dhcp.@dnsmasq[0] | grep "\.server=" | awk -F '=' '{print $2}' | sed "s/'//g" | tr ' ' '\n' | grep -v "\/" | head -2 | sed ':label;N;s/\n/,/;b label') [ -z "${DEFAULT_DNS}" ] && [ "$(echo $ISP_DNS | tr ' ' '\n' | wc -l)" -le 2 ] && DEFAULT_DNS=$(echo -n $ISP_DNS | tr ' ' '\n' | head -2 | tr '\n' ',' | sed 's/,$//') LOCAL_DNS="${DEFAULT_DNS:-119.29.29.29,223.5.5.5}" IPT_APPEND_DNS=${LOCAL_DNS} DNSMASQ_CONF_DIR=/tmp/dnsmasq.d DEFAULT_DNSMASQ_CFGID="$(uci -q show "dhcp.@dnsmasq[0]" | awk 'NR==1 {split($0, conf, /[.=]/); print conf[2]}')" if [ -f "/tmp/etc/dnsmasq.conf.$DEFAULT_DNSMASQ_CFGID" ]; then DNSMASQ_CONF_DIR="$(awk -F '=' '/^conf-dir=/ {print $2}' "/tmp/etc/dnsmasq.conf.$DEFAULT_DNSMASQ_CFGID")" if [ -n "$DNSMASQ_CONF_DIR" ]; then DNSMASQ_CONF_DIR=${DNSMASQ_CONF_DIR%*/} else DNSMASQ_CONF_DIR="/tmp/dnsmasq.d" fi fi set_cache_var GLOBAL_DNSMASQ_CONF ${DNSMASQ_CONF_DIR}/dnsmasq-${CONFIG}.conf set_cache_var GLOBAL_DNSMASQ_CONF_PATH ${GLOBAL_ACL_PATH}/dnsmasq.d } arg1=$1 shift case $arg1 in add_ip2route) add_ip2route $@ ;; echolog) echolog $@ ;; get_new_port) get_new_port $@ ;; get_cache_var) get_cache_var $@ ;; set_cache_var) set_cache_var $@ ;; run_socks) run_socks $@ ;; socks_node_switch) socks_node_switch $@ ;; start) start ;; stop) stop ;; esac
294coder/Efficient-MIF
22,774
utils/train_utils.py
from pathlib import Path import sys sys.path.append("./") import zipfile from io import BytesIO from pathlib import Path from PIL import Image import PIL.Image as Image from contextlib import contextmanager from collections import OrderedDict from typing import Union, TYPE_CHECKING import h5py import numpy as np import torch from torch.utils.data import DataLoader from utils import h5py_to_dict, NameSpace, easy_logger if TYPE_CHECKING: from accelerate import Accelerator def get_eval_dataset(args, logger=None): from task_datasets.WV3 import WV3Datasets from task_datasets.GF2 import GF2Datasets from task_datasets.HISR import HISRDatasets from task_datasets.TNO import TNODataset from task_datasets.RoadScene import RoadSceneDataset from task_datasets.LLVIP import LLVIPDALIPipeLoader from task_datasets.MSRS import MSRSDatasets from task_datasets.M3FD import M3FDDALIPipeLoader from task_datasets.MedHarvard import MedHarvardDataset from task_datasets.SICE import SICEDataset logger = easy_logger(func_name='get_eval_dataset') val_ds, val_dl = None, None logger.info(f"use dataset: {args.dataset} on VIS-IR fusion task") # 1. vis-ir image fusion (without gt) if args.dataset == "roadscene": val_ds = RoadSceneDataset( args.path.base_dir, "test", no_split=True, get_name=True ) elif args.dataset == "tno": val_ds = TNODataset( args.path.base_dir, "test", aug_prob=0.0, no_split=True, get_name=True ) elif args.dataset == "msrs": val_ds = MSRSDatasets( args.path.base_dir, mode=args.dataset_mode, # or 'test'/'detection' transform_ratio=0.0, get_name=True, reduce_label=args.reduce_label, ) elif args.dataset == "llvip": val_dl = LLVIPDALIPipeLoader( args.path.base_dir, "test", batch_size=args.val_bs, device=args.device, shuffle=False, with_mask=True, get_name=True, reduce_label=args.reduce_label, ) elif args.dataset == "m3fd": val_dl = M3FDDALIPipeLoader( args.path.base_dir, "test", batch_size=args.val_bs, device=args.device, shuffle=False, with_mask=True, get_name=True, reduce_label=args.reduce_label, ) elif args.dataset == "med_harvard": val_ds = MedHarvardDataset( args.path.base_dir, mode="test", device=args.device, data_source="xmu", get_name=True, task="SPECT-MRI", ) elif args.dataset == "sice": val_ds = SICEDataset( data_dir=args.path.base_dir, mode="test", transformer_ratio=0.0, only_y=args.only_y, get_name=True, ) ## 2. sharpening datasets (with gt) elif args.dataset in [ "wv3", "qb", "gf2", "cave_x4", "harvard_x4", "cave_x8", "harvard_x8", "hisi-houston", "pavia", "chikusei", "botswana", ]: # the dataset has already splitted logger.info(f"use dataset: {args.dataset} on pansharpening/HISR task") # FIXME: 需要兼顾老代码(只有trian_path和val_path)的情况 if hasattr(args.path, "val_path"): # 旧代码:手动切换数据集路径 val_path = args.path.val_path else: _args_path_keys = list(args.path.__dict__.keys()) for k in _args_path_keys: if args.dataset in k: val_path = getattr(args.path, f"{args.dataset}_val_path") assert val_path is not None, "val_path should not be None" if val_path is not None: assert val_path.endswith(".h5"), 'val_path should end with ".h5"' h5_val = h5py.File(val_path) # 1. parsharpening if args.dataset in ["wv3", "qb"]: d_val = h5py_to_dict(h5_val) val_ds = WV3Datasets(d_val, hp=args.hp, aug_prob=0.0) elif args.dataset == "gf2": d_val = h5py_to_dict(h5_val) val_ds = GF2Datasets(d_val, hp=args.hp, aug_prob=0.0) # 2. hyperspectral image fusion elif ( args.dataset[:4] == "cave" or args.dataset[:7] == "harvard" or args.dataset[:8] == "chikusei" or args.dataset[:5] == "pavia" or args.dataset[:8] == "botswana" ): keys = ["LRHSI", "HSI_up", "RGB", "GT"] if args.dataset.split("-")[-1] == "houston": from einops import rearrange # to avoid unpicklable error def permute_fn(x): return rearrange(x, "b h w c -> b c h w") dataset_fn = permute_fn else: dataset_fn = None d_val = h5py_to_dict(h5_val, keys) val_ds = HISRDatasets( d_val, aug_prob=args.aug_probs[1], dataset_fn=dataset_fn ) # del h5_train, h5_val else: raise NotImplementedError(f"not support dataset {args.dataset}") return val_ds, val_dl def get_fusion_dataset( args: NameSpace, accelerator: "Accelerator", device: Union[str, torch.device] ): logger = easy_logger() train_ds, val_ds, train_dl, val_dl = None, None, None, None if args.dataset in [ "flir", "tno", "roadscene_tno_joint", "vis_ir_joint", "msrs", "llvip", "med_harvard", "m3fd", "sice", "mefb", ]: args.task = "fusion" args.has_gt = False args.path.base_dir = getattr(args.path, f"{args.dataset}_base_dir") if args.dataset == "roadscene": from task_datasets.RoadScene import RoadSceneDataset train_ds = RoadSceneDataset(args.path.base_dir, "train") val_ds = RoadSceneDataset(args.path.base_dir, "test") elif args.dataset in ["tno", "roadscene_tno_joint"]: from task_datasets.TNO import TNODataset train_ds = TNODataset( args.path.base_dir, "train", aug_prob=args.aug_probs[0], duplicate_vis_channel=True, ) val_ds = TNODataset( args.path.base_dir, "test", aug_prob=args.aug_probs[1], no_split=True, duplicate_vis_channel=True, ) elif args.dataset == "msrs": from task_datasets.MSRS import MSRSDatasets train_ds = MSRSDatasets( args.path.base_dir, "train", transform_ratio=args.aug_probs[0], output_size=args.fusion_crop_size, n_proc_load=1, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, # args.only_y ) val_ds = MSRSDatasets( args.path.base_dir, "test", transform_ratio=args.aug_probs[1], output_size=args.fusion_crop_size, n_proc_load=1, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, # args.only_y ) elif args.dataset == "llvip": from task_datasets.LLVIP import LLVIPDALIPipeLoader # We use DALI pipeline to accelerate the data loading process train_dl = LLVIPDALIPipeLoader( args.path.base_dir, "train", batch_size=args.train_bs, output_size=args.fusion_crop_size, device=accelerator.device, num_shards=accelerator.num_processes, shard_id=accelerator.process_index, shuffle=True, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, # args.only_y ) val_dl = LLVIPDALIPipeLoader( args.path.base_dir, "test", batch_size=args.val_bs, device=accelerator.device, fast_eval_n_samples=args.fast_eval_n_samples, num_shards=accelerator.num_processes, shard_id=accelerator.process_index, shuffle=True, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, # args.only_y ) elif args.dataset == "m3fd": from task_datasets import M3FDDALIPipeLoader train_dl = M3FDDALIPipeLoader( args.path.base_dir, "train", batch_size=args.train_bs, output_size=args.fusion_crop_size, device=accelerator.device, num_shards=accelerator.num_processes, shard_id=accelerator.process_index, shuffle=True, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, # args.only_y ) val_dl = M3FDDALIPipeLoader( args.path.base_dir, "test", batch_size=args.val_bs, device=accelerator.device, fast_eval_n_samples=args.fast_eval_n_samples, num_shards=accelerator.num_processes, shard_id=accelerator.process_index, shuffle=True, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, # args.only_y ) elif args.dataset == "vis_ir_joint": from task_datasets import VISIRJointGenericLoader train_dl = VISIRJointGenericLoader( vars(args.path.base_dir), mode="train", batch_size=args.train_bs, output_size=args.fusion_crop_size, device=accelerator.device, shuffle_in_dataset=True, num_shards=accelerator.num_processes, shard_id=accelerator.process_index, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, ) val_dl = VISIRJointGenericLoader( ## only test msrs and roadscene_tno_joint dataset {'msrs': args.path.base_dir['msrs'], 'roadscene_tno_joint': args.path.base_dir['roadscene_tno_joint']}, mode="test", output_size=224, # enforce the different size images to be the same size batch_size=args.val_bs, device=accelerator.device, shuffle_in_dataset=False, fast_eval_n_samples=30, num_shards=accelerator.num_processes, shard_id=accelerator.process_index, reduce_label=args.datasets_cfg.reduce_label, only_y_component=False, ) elif args.dataset == "med_harvard": from task_datasets.MedHarvard import MedHarvardDataset if getattr(args, "datasets_cfg", None): task = args.datasets_cfg.med_harvard.task else: task = None train_ds = MedHarvardDataset( args.path.base_dir, mode="train", device=device, data_source="xmu", transform_ratio=args.aug_probs[0], task=task, ) val_ds = MedHarvardDataset( args.path.base_dir, mode="test", device=device, data_source="xmu", task=task, ) assert args.num_worker == 0, "num_worker should be 0 for MedHarvard dataset" elif args.dataset == "sice": from task_datasets.SICE import SICEDataset train_ds = SICEDataset( data_dir=args.path.base_dir, mode="train", transformer_ratio=args.aug_probs[0], output_size=args.fusion_crop_size, only_y=args.only_y, ) val_ds = SICEDataset( data_dir=args.path.base_dir, mode="train", transformer_ratio=args.aug_probs[0], output_size=args.fusion_crop_size, only_y=args.only_y, ) args.has_gt = True else: raise NotImplementedError(f"not support dataset {args.dataset}") elif args.dataset in [ "wv3", "qb", "gf2", "cave_x4", "harvard_x4", "cave_x8", "harvard_x8", "hisi-houston", ]: args.task = "sharpening" # the dataset has already splitted # FIXME: 需要兼顾老代码(只有trian_path和val_path)的情况 if hasattr(args.path, "train_path") and hasattr(args.path, "val_path"): # 旧代码:手动切换数据集路径 train_path = args.path.train_path val_path = args.path.val_path else: _args_path_keys = list(args.path.__dict__.keys()) for k in _args_path_keys: if args.dataset in k: train_path = getattr(args.path, f"{args.dataset}_train_path") val_path = getattr(args.path, f"{args.dataset}_val_path") assert ( train_path is not None and val_path is not None ), "train_path and val_path should not be None" h5_train, h5_val = ( h5py.File(train_path), h5py.File(val_path), ) if args.dataset in ["wv3", "qb"]: from task_datasets.WV3 import WV3Datasets, make_datasets d_train, d_val = h5py_to_dict(h5_train), h5py_to_dict(h5_val) train_ds, val_ds = ( WV3Datasets(d_train, aug_prob=args.aug_probs[0]), WV3Datasets(d_val, aug_prob=args.aug_probs[1]), ) elif args.dataset == "gf2": from task_datasets.GF2 import GF2Datasets d_train, d_val = h5py_to_dict(h5_train), h5py_to_dict(h5_val) train_ds, val_ds = ( GF2Datasets(d_train, aug_prob=args.aug_probs[0]), GF2Datasets(d_val, aug_prob=args.aug_probs[1]), ) elif args.dataset[:4] == "cave" or args.dataset[:7] == "harvard": from task_datasets.HISR import HISRDatasets keys = ["LRHSI", "HSI_up", "RGB", "GT"] if args.dataset.split("-")[-1] == "houston": from einops import rearrange def permute_fn(x): return rearrange(x, "b h w c -> b c h w") dataset_fn = permute_fn else: dataset_fn = None d_train, d_val = ( h5py_to_dict(h5_train, keys), h5py_to_dict(h5_val, keys), ) train_ds = HISRDatasets( d_train, aug_prob=args.aug_probs[0], dataset_fn=dataset_fn ) val_ds = HISRDatasets( d_val, aug_prob=args.aug_probs[1], dataset_fn=dataset_fn ) # del h5_train, h5_val else: raise NotImplementedError(f"not support dataset {args.dataset}") train_sampler, val_sampler = None, None if train_dl is None: train_dl = DataLoader( train_ds, args.train_bs, num_workers=args.num_worker, sampler=train_sampler, prefetch_factor=8 if args.num_worker > 0 else None, pin_memory=False, shuffle=args.shuffle if not args.ddp else None, drop_last=True if args.shuffle else False, ) if val_dl is None: val_dl = DataLoader( val_ds, args.val_bs, # assert bs is 1, when using PatchMergeModule num_workers=0, sampler=val_sampler, pin_memory=False, shuffle=args.shuffle if not args.ddp else None, drop_last=False, ) return train_ds, train_dl, val_ds, val_dl def set_ema_model_params_with_keys(ema_model_params: "dict[str, list[torch.Tensor] | int | float]", keys: "list[str]", keys_set: list[str]=['shadow_params']): """set ema model parameters with keys Args: ema_model_params (dict[str, list[torch.Tensor] | int | float]): ema model parameters keys (list[str]): keys Returns: dict: ema model parameters with keys """ logger = easy_logger() if not isinstance(keys, list): keys = list(keys) ema_model_params_with_keys = OrderedDict() for k in ema_model_params.keys(): if k in keys_set and k in ema_model_params: logger.info(f'set ema_model {k} params with keys') params = ema_model_params[k] assert params is not None assert len(params) == len(keys), "ema_model_params and keys should have the same length" _params = OrderedDict() for mk, p in zip(keys, params): _params[mk] = p ema_model_params_with_keys[k] = _params elif k not in keys_set and k in ema_model_params: ema_model_params_with_keys[k] = ema_model_params[k] return ema_model_params_with_keys def run_once(abled=True): def _inner(func): def _wrapper(*args, **kwargs): nonlocal abled if not abled: return None else: outs = func(*args, **kwargs) abled = False return outs return _wrapper return _inner def sanity_check(func: callable): @run_once() def _inner(*args, **kwargs): return func(*args, **kwargs) return _inner @contextmanager def save_imgs_in_zip( zipfile_name: str, mode="w", verbose: bool = False, save_file_ext: str = "jpeg" ): """save images to a zip file Args: zipfile_name (str): zip filename mode (str, optional): mode to write in. Defaults to "w". verbose (bool, optional): print out. Defaults to False. save_file_ext (str, optional): image extension in the zip file. Defaults to "jpeg". Yields: callable: a function to save image Examples:: with save_imgs_in_zip('zip_file.zip') as add_image: img, img_name = get_img() add_image(img, img_name) :ref: `add_image` """ logger = easy_logger() # save_file_ext = save_file_ext.upper() zf = zipfile.ZipFile( zipfile_name, mode=mode, compression=zipfile.ZIP_DEFLATED, compresslevel=9 ) bytes_io = BytesIO() # jpg compression _jpg_quality = 100 # 95 if save_file_ext in ["jpeg", "jpg", "JPG", "JPEG"] else 100 try: logger.info(f"zip file will be saved at {zipfile_name}") def to_bytes(image_data, image_name): batched_image_bytes = [] if image_data.ndim == 4: # batched rgb images assert isinstance(image_name, list), "image_name should be a list" assert image_data.shape[0] == len( image_name ), "image_name should have the same length as image_data" for img in image_data: # [b, h, w, c] Image.fromarray(img).save( bytes_io, format=save_file_ext, quality=_jpg_quality ) batched_image_bytes.append(bytes_io.getvalue()) elif image_data.ndim == 3: if image_data.shape[-1] == 1: # gray image # [h, w, 1] Image.fromarray(image_data[..., 0]).save( bytes_io, format=save_file_ext, quality=_jpg_quality ) image_data = bytes_io.getvalue() elif image_data.shape[-1] == 3: Image.fromarray(image_data).save( bytes_io, format=save_file_ext, quality=_jpg_quality ) image_data = bytes_io.getvalue() else: raise ValueError( f"image_data shape {image_data.shape} not supported" ) elif image_data.ndim == 2: # gray image # [h, w] Image.fromarray(image_data).save( bytes_io, format=save_file_ext, quality=_jpg_quality ) image_data = bytes_io.getvalue() return image_data, batched_image_bytes def add_image( image_data: "Image.Image | np.ndarray | torch.Tensor | bytes", image_name: "Union[str, list[str]]", ): """add image to the zipfile Args: image_data (Image.Image | np.ndarray | torch.Tensor | bytes): can be Image.Image, np.ndarray, torch.Tensor, bytes, shape should be [b, h, w, c], [h, w, c], [h, w, 1] image_name (str | list[str]): saved image names """ # to bytes batched_image_bytes = None if isinstance(image_data, Image.Image): image_data.save(bytes_io, format=save_file_ext) bytes = bytes_io.getvalue() elif isinstance(image_data, np.ndarray): bytes, batched_image_bytes = to_bytes(image_data, image_name) elif isinstance(image_data, torch.Tensor): image_data = image_data.detach().cpu().numpy() bytes, batched_image_bytes = to_bytes(image_data, image_name) else: raise ValueError(f"image_data type {type(image_data)} not supported") # saving to zip file if batched_image_bytes is not None: for i, img_bytes in enumerate(batched_image_bytes): zf.writestr(image_name[i], img_bytes) else: zf.writestr(image_name, bytes) if verbose: logger.info(f"add image {image_name} to zip file") bytes_io.seek(0) bytes_io.truncate() yield add_image except Exception as e: if verbose: logger.error(e, raise_error=True) raise e finally: if verbose: logger.info(f"zip file saved at {zipfile_name}, zipfile close") zf.close() bytes_io.close()
2977094657/DidaAPI
2,118
frontend/docs/api/pomodoros/focus-distribution.md
# 获取专注详情分布 获取指定日期范围内的专注时长分布统计,包括按项目、标签、任务分布的专注时长。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/pomodoros/statistics/dist/{start_date}/{end_date}` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单官方API ## 请求参数 ### 路径参数 | 参数名 | 类型 | 必填 | 描述 | 示例 | |--------|------|------|------|------| | start_date | string | 是 | 开始日期,格式: YYYYMMDD | 20231201 | | end_date | string | 是 | 结束日期,格式: YYYYMMDD | 20231207 | ## 完整请求示例 ```http GET https://api.dida365.com/api/v2/pomodoros/statistics/dist/20231201/20231207 HTTP/1.1 Host: api.dida365.com Cookie: t=string; _csrf_token=string X-CSRFToken: string ``` ## 响应格式 ### 成功响应 **状态码**: `200 OK` **响应体**: ```json { "projectDurations": { "project_id_1": { "duration": 3600, "name": "工作项目", "color": "#FF5722" }, "project_id_2": { "duration": 1800, "name": "学习项目", "color": "#2196F3" } }, "tagDurations": { "tag_id_1": { "duration": 2400, "name": "编程", "color": "#4CAF50" }, "tag_id_2": { "duration": 1200, "name": "阅读", "color": "#FF9800" } }, "taskDurations": { "task_id_1": { "duration": 1800, "title": "完成API文档", "projectId": "project_id_1" }, "task_id_2": { "duration": 1200, "title": "学习新技术", "projectId": "project_id_2" } } } ``` ## 响应字段说明 ### projectDurations (按项目分布) - **类型**: Object - **描述**: 按项目分组的专注时长统计 - **字段说明**: - `duration`: 专注时长(秒) - `name`: 项目名称 - `color`: 项目颜色 ### tagDurations (按标签分布) - **类型**: Object - **描述**: 按标签分组的专注时长统计 - **字段说明**: - `duration`: 专注时长(秒) - `name`: 标签名称 - `color`: 标签颜色 ### taskDurations (按任务分布) - **类型**: Object - **描述**: 按任务分组的专注时长统计 - **字段说明**: - `duration`: 专注时长(秒) - `title`: 任务标题 - `projectId`: 所属项目ID ## 使用说明 1. **日期格式**: 必须使用 YYYYMMDD 格式 2. **时长单位**: 所有时长都以秒为单位 3. **数据范围**: 返回指定日期范围内的专注分布统计 4. **认证要求**: 需要有效的登录会话 ## 相关接口 - [获取专注记录时间线](./focus-timeline.md) - 获取详细的专注记录 - [获取专注趋势热力图](./focus-heatmap.md) - 获取专注趋势数据 - [获取专注时间分布](./focus-time-distribution.md) - 获取时间段分布 - [获取专注时间按小时分布](./focus-hour-distribution.md) - 获取小时分布
2977094657/DidaAPI
910
frontend/docs/api/pomodoros/focus-time-distribution.md
# 获取专注时间分布 获取指定日期范围内按时间段分布的专注数据。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/pomodoros/statistics/dist/clockByDay/{start_date}/{end_date}` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 | 参数名 | 类型 | 必填 | 说明 | 示例 | |--------|------|------|------|------| | start_date | string | 是 | 开始日期,格式: YYYYMMDD | 20250526 | | end_date | string | 是 | 结束日期,格式: YYYYMMDD | 20250601 | ## 响应格式 ### 成功响应 ```json [ { "day": "20250526", "timezone": "Asia/Shanghai" }, { "timeDurations": { "15": 60, "22": 60, "23": 60 }, "day": "20250601", "timezone": "Asia/Shanghai" } ] ``` ### 响应字段说明 | 字段 | 类型 | 说明 | |------|------|------| | day | string | 日期(YYYYMMDD格式) | | timezone | string | 时区 | | timeDurations | object | 时间段专注分布(可选字段) | | timeDurations.{hour} | number | 指定小时的专注时长(分钟),hour为24小时制 |
294coder/Efficient-MIF
11,517
utils/inference_helper_func.py
# GPL License # Copyright (C) 2024 , UESTC # All Rights Reserved # # @Time : 2021/10/15 17:53 # @Author : Zihan Cao, Xiao Wu from functools import partial import inspect from typing import Tuple, Optional import einops import torch import torch.nn as nn import torch.nn.functional as F import math from torch.utils.data import DataLoader from torch import Tensor, nn from tqdm import tqdm import json import re import glob from .visualize import viz_batch, res_image from .metric_sharpening import AnalysisPanAcc from typing import TYPE_CHECKING if TYPE_CHECKING: from model.base_model import BaseModel def has_patch_merge_model(model: "nn.Module | BaseModel"): return (hasattr(model, '_patch_merge_model')) or (hasattr(model, 'patch_merge_model')) def patch_merge_in_val_step(model: "nn.Module | BaseModel"): return 'patch_merge' in list(inspect.signature(model.val_step).parameters.keys()) def has_patch_merge_model(model: "nn.Module | BaseModel"): return (hasattr(model, '_patch_merge_model')) or (hasattr(model, 'patch_merge_model')) def patch_merge_in_val_step(model): return 'patch_merge' in list(inspect.signature(model.val_step).parameters.keys()) # callback function def basic_callback(model: "BaseModel", iter_idx: int): from utils import get_local assert get_local().cache is not None and get_local.is_activate cache = get_local().cache attns = cache['_forward_implem'] get_local.clear() ############ Main inference function ############ @torch.no_grad() @torch.inference_mode() def unref_for_loop(model, dl: DataLoader, device, *, split_patch=False, feature_callback: callable=None, **patch_merge_module_kwargs): from model.base_model import PatchMergeModule all_sr = [] try: spa_size = tuple(dl.dataset.lms.shape[-2:]) except AttributeError: spa_size = tuple(dl.dataset.rgb.shape[-2:]) inference_bar = tqdm(enumerate(dl, 1), dynamic_ncols=True, total=len(dl)) analysis = AnalysisPanAcc(ratio=patch_merge_module_kwargs.get('ergas_ratio', 4), ref=False, sensor=patch_merge_module_kwargs.get('sensor', 'DEFAULT'), default_max_value=patch_merge_module_kwargs.get('default_max_value', None)) if split_patch: # check if has the patch merge model if not (has_patch_merge_model(model) or patch_merge_in_val_step(model)): # assert bs == 1, 'batch size should be 1' # warp the model into PatchMergeModule model = PatchMergeModule(net=model, device=device, **patch_merge_module_kwargs) for i, (pan, ms, lms) in inference_bar: pan, ms, lms = pan.to(device).float(), ms.to(device).float(), lms.to(device).float() # split the image into several patches to avoid gpu OOM if split_patch: input = (ms, lms, pan) if hasattr(model, 'forward_chop'): sr = model.forward_chop(*input)[0] elif patch_merge_in_val_step(model): sr = model.val_step(*input, patch_merge=True) else: raise NotImplemented('model should have @forward_chop or patch_merge arg in @val_step') else: if patch_merge_in_val_step(model): sr = model.val_step(ms, lms, pan, False) else: sr = model.val_step(ms, lms, pan) sr = sr.clip(0, 1) sr1 = sr.detach().cpu().numpy() all_sr.append(sr1) # analysis(sr, ms, lms, pan) viz_batch(sr.detach().cpu(), suffix='sr', start_index=i, base_path='visualized_img/img_shows') viz_batch(ms.detach().cpu(), suffix='ms', start_index=i, base_path='visualized_img/img_shows') viz_batch(pan.detach().cpu(), suffix='pan', start_index=i, base_path='visualized_img/img_shows') if feature_callback is not None: feature_callback(model, i) print(analysis.print_str()) return all_sr @torch.no_grad() @torch.inference_mode() def ref_for_loop(model, dl, device, *, split_patch=False, ergas_ratio=4, residual_exaggerate_ratio=100, feature_callback: callable=None, **patch_merge_module_kwargs): from model.base_model import PatchMergeModule analysis = AnalysisPanAcc(ergas_ratio) all_sr = [] inference_bar = tqdm(enumerate(dl, 1), dynamic_ncols=True, total=len(dl)) if not (has_patch_merge_model(model) or patch_merge_in_val_step(model)): model = PatchMergeModule(net=model, device=device, **patch_merge_module_kwargs) for i, (pan, ms, lms, gt) in inference_bar: pan, ms, lms, gt = pan.to(device).float(), ms.to(device).float(), lms.to(device).float(), gt.to(device).float() # split the image into several patches to avoid gpu OOM if split_patch: input = (ms, lms, pan) if hasattr(model, 'forward_chop'): sr = model.forward_chop(*input)[0] elif patch_merge_in_val_step(model): sr = model.val_step(*input, patch_merge=split_patch) else: raise NotImplemented('model should have @forward_chop or patch_merge arg in @val_step') else: if patch_merge_in_val_step(model): sr = model.val_step(ms, lms, pan, False) else: sr = model.val_step(ms, lms, pan) if feature_callback is not None: feature_callback(model, i) sr = sr.clip(0, 1) sr1 = sr.detach().cpu().numpy() all_sr.append(sr1) analysis(gt, sr) res = res_image(gt, sr, exaggerate_ratio=residual_exaggerate_ratio) viz_batch(sr.detach().cpu(), suffix='sr', start_index=i, base_path='visualized_img/img_shows') viz_batch(gt.detach().cpu(), suffix='gt', start_index=i, base_path='visualized_img/img_shows') viz_batch(ms.detach().cpu(), suffix='ms', start_index=i, base_path='visualized_img/img_shows') viz_batch(pan.detach().cpu(), suffix='pan', start_index=i, base_path='visualized_img/img_shows') viz_batch(res.detach().cpu(), suffix='residual', start_index=i, base_path='visualized_img/img_shows') print(analysis.print_str()) return all_sr def find_data_path(dataset_type, full_res): if dataset_type == "wv3": if not full_res: path = "/volsparse1/dataset/PanCollection/test_data/test_wv3_multiExm1.h5" else: path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_wv3_OrigScale_multiExm1.h5" elif dataset_type == "cave": path = "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5" elif dataset_type == "cave_x8": path = "/volsparse1/dataset/HISR/cave_x8/test_cave(with_up)x8_rgb.h5" elif dataset_type == "harvard": path = "/Data2/ShangqiDeng/data/HSI/harvard_x4/test_harvard(with_up)x4_rgb200.h5" elif dataset_type == "harvard_x8": path = "/volsparse1/dataset/HISR/harvard_x8/test_harvard(with_up)x8_rgb.h5" elif dataset_type == "gf5": if not full_res: path = "/Data2/ZiHanCao/datasets/pansharpening/GF5-GF1/tap23/test_GF5_GF1_23tap_new.h5" else: path = "/Data2/ZiHanCao/datasets/pansharpening/GF5-GF1/tap23/test_GF5_GF1_OrigScale.h5" elif dataset_type == "gf": if not full_res: path = "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5" else: path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_gf2_OrigScale_multiExm1.h5" elif dataset_type == "qb": if not full_res: path = "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5" else: path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_qb_OrigScale_multiExm1.h5" elif dataset_type == "wv2": if not full_res: path = "/Data2/ZiHanCao/datasets/pansharpening/wv2/reduced_examples/test_wv2_multiExm1.h5" else: path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_wv2_OrigScale_multiExm1.h5" elif dataset_type == "roadscene": path = "/Data2/ZiHanCao/datasets/RoadSceneFusion_1" elif dataset_type == "tno": path = "/Data2/ZiHanCao/datasets/TNO" else: raise NotImplementedError("not exists {} dataset".format(dataset_type)) return path def find_key_args_in_log(arch, sub_arch, datasets, weight_path): # handle weight_path slash_with_id = re.findall(r'_[a-zA-Z0-9-]{8}(?=\.pth|_)', weight_path)[-1] run_id = slash_with_id[1:] if sub_arch is not None and sub_arch != '': sub_arch = '_' + sub_arch else: sub_arch = '' _log_path = f'log_file/{arch}{sub_arch}/{datasets}/*{run_id}*/config.json' log_path = glob.glob(_log_path) if len(log_path) != 1: raise RuntimeError(f'>>> log file: {_log_path} not exists!') print(f'>>> found run id: {log_path[0]} config') args = json.loads(''.join(open(log_path[0], 'r').readlines())) return args def crop_inference(model: "BaseModel", xs: Tuple[Tensor, Tensor, Tensor], crop_size: Tuple[int] = (16, 64, 64), stride: Tuple[int] = (8, 32, 32)): # only support CAVE dataset # input shape: 128, 512, 512 # xs: (hsi_lr, hsi_up, rgb) torch.backends.cudnn.enable = True torch.backends.cudnn.benchmark = True # preprocessing crop_xs = [] ncols = [] bs, out_c, _, _ = xs[0].shape _, _, out_h, out_w = xs[-1].shape for i in range(len(xs)): x = xs[i] _, c, h, _ = x.shape # assume h equals w crop = crop_size[i] s = stride[i] ncol = (h - crop) // s ncols.append(ncol) crop_x = F.unfold(x, crop, stride=s) crop_x = einops.rearrange(crop_x, 'b (c k l) m -> m b c k l', k=crop, l=crop, c=c) crop_xs.append(crop_x) # model inference model.eval() out = [] for i in range(crop_xs[0].size(0)): input = [crop_xs[j][i].cuda(0) for j in range(len(xs))] out.append(model.val_step(*input).detach().cpu()) # [bs * 225, 31, 64, 64] del input torch.cuda.empty_cache() # input: 255*[b, 31, 64, 64] out = torch.cat(out, dim=0) # postprocessing out = einops.rearrange(out, '(m b) c k l -> b (c k l) m', b=bs, k=crop_size[-1], l=crop_size[-1], c=out_c) output = F.fold(out, output_size=(out_h, out_w), kernel_size=(crop_size[-1], crop_size[-1]), dilation=1, padding=0, stride=(stride[-1], stride[-1])) return output if __name__ == '__main__': from model.dcformer_reduce import DCFormer_Reduce model = DCFormer_Reduce(8, 'C').cuda(0) ms = torch.randn(1, 8, 128, 128) interp_ms = F.interpolate(ms, size=512) lms = torch.randn(1, 8, 512, 512) pan = torch.randn(1, 1, 512, 512) expand_pan = pan.expand(-1, 8, -1, -1) # print(model.val_step(ms, lms, pan).shape) print(crop_inference(model, xs=(ms, lms, pan)).shape)
297854895/vue-tsx-admin
2,780
.eslintrc.js
module.exports = { root: true, env: { node: true }, extends: [ "plugin:vue/essential", "@vue/typescript" ], rules: { "no-console": process.env.NODE_ENV === "production" ? "error" : "off", "no-debugger": process.env.NODE_ENV === "production" ? "error" : "off", "quotes": [2, "single"], //单引号 'new-cap': 'off', 'no-continue': 'off', 'no-empty': 'on', 'experimentalDecorators': 'off', 'consistent-return': 'off', 'no-plusplus': 'off', 'guard-for-in': 'off', 'prefer-destructuring': 'off', 'prefer-promise-reject-errors': 'off', 'camelcase': 'off', 'eslint-import-resolver-typescript': { 'extensions': ['.ts', '.tsx', '.d.ts'] }, 'space-infix-ops': 'off', 'import/prefer-default-export': 'off', 'no-restricted-syntax': 0, 'no-restricted-globals': 'off', 'global-require': 'off', 'import/no-dynamic-require': 'off', 'import/no-unresolved': 'off', 'import/newline-after-import': 'off', 'import/extensions': 'off', 'no-param-reassign': 'off', 'class-methods-use-this': 'off', 'linebreak-style': 'off', 'no-console': 'off', 'no-debugger': 'off', "no-console": 0, //不禁用console // "no-debugger": 2, //禁用debugger "no-var": 0, //对var警告 "semi": 0, //不强制使用分号 "no-irregular-whitespace": 0, //不规则的空白不允许 "no-trailing-spaces": 1, //一行结束后面有空格就发出警告 "eol-last": 0, //文件以单一的换行符结束 // "no-unused-vars": [2, {"vars": "local", "args": "after-used"}], //不能有声明后未被使用的变量或参数 "no-underscore-dangle": 0, //标识符不能以_开头或结尾 "no-alert": 2, //禁止使用alert confirm prompt "no-lone-blocks": 0, //禁止不必要的嵌套块 "no-class-assign": 2, //禁止给类赋值 "no-cond-assign": 2, //禁止在条件表达式中使用赋值语句 "no-const-assign": 2, //禁止修改const声明的变量 "no-delete-var": 2, //不能对var声明的变量使用delete操作符 "no-dupe-keys": 2, //在创建对象字面量时不允许键重复 "no-duplicate-case": 2, //switch中的case标签不能重复 "no-dupe-args": 2, //函数参数不能重复 "no-empty": 2, //块语句中的内容不能为空 "no-func-assign": 2, //禁止重复的函数声明 "no-invalid-this": 0, //禁止无效的this,只能用在构造器,类,对象字面量 "no-redeclare": 2, //禁止重复声明变量 "no-spaced-func": 2, //函数调用时 函数名与()之间不能有空格 "no-this-before-super": 0, //在调用super()之前不能使用this或super "no-undef": 2, //不能有未定义的变量 "no-use-before-define": 2, //未定义前不能使用 "jsx-quotes": [2, "prefer-double"], //强制在JSX属性(jsx-quotes)中一致使用双引号 'vue/attribute-hyphenation': [ 'error', 'always' ], 'vue/html-end-tags': 'error', 'vue/html-indent': [ 'error', 2 ], 'vue/html-self-closing': 'error', 'vue/require-default-prop': 'error', 'vue/require-prop-types': 'error', 'vue/attributes-order': 'error', 'vue/order-in-components': 'error' }, parserOptions: { parser: "@typescript-eslint/parser" } };
281677160/openwrt-package
27,599
luci-app-passwall/root/usr/share/passwall/rules/chnroute6
2001:250::/30 2001:254::/31 2001:256:100::/48 2001:4510:400::/40 2001:678:10d0::/48 2001:678:120::/48 2001:67c:ebc::/48 2001:7fa:5::/48 2001:c68::/32 2001:cc0::/32 2001:da8::/32 2001:daa:1::/48 2001:daa:2::/47 2001:daa:4::/47 2001:daa:6::/48 2001:dc7::/32 2001:dd8:1::/48 2001:dd9::/48 2001:df0:ac40::/48 2001:df1:61c0::/48 2001:df3:3a80::/48 2001:df3:8b80::/48 2001:df7:1480::/48 2400:1160::/32 2400:3200::/32 2400:5280:f803::/48 2400:5400:10::/48 2400:5a60:100::/48 2400:5a60:2::/48 2400:5f60::/32 2400:6000::/32 2400:6460:300::/40 2400:6460:500::/40 2400:6460::/39 2400:6600::/32 2400:6e60:1301::/48 2400:7100::/32 2400:73e0::/32 2400:75aa::/32 2400:7bc0:20::/43 2400:7fc0:220::/44 2400:7fc0:240::/44 2400:7fc0:2a0::/44 2400:7fc0:2c0::/44 2400:7fc0:4000::/40 2400:7fc0:4100::/48 2400:7fc0:6000::/40 2400:7fc0:8000::/36 2400:7fc0::/40 2400:7fc0:a000::/36 2400:7fc0:bb00::/40 2400:7fc0:c000::/36 2400:8200::/32 2400:87c0::/32 2400:89c0:1010::/44 2400:89c0:1020::/44 2400:89c0:1050::/46 2400:89c0:1130::/44 2400:89c0:1150::/48 2400:89c0:2100::/48 2400:89c0:2200::/48 2400:89c0:3010::/44 2400:89c0:6000::/48 2400:89c0:6100::/48 2400:8fc0::/32 2400:9020:f010::/46 2400:9340::/32 2400:9380:8001::/48 2400:9380:8003::/48 2400:9380:8021::/48 2400:9380:8040::/48 2400:9380:8140::/48 2400:9380:8201::/48 2400:9380:8301::/48 2400:9380:9001::/48 2400:9380:9002::/48 2400:9380:9005::/48 2400:9380:9009::/48 2400:9380:900a::/48 2400:9380:9020::/47 2400:9380:9040::/47 2400:9380:9050::/47 2400:9380:9060::/48 2400:9380:9071::/48 2400:9380:9080::/47 2400:9380:90b0::/45 2400:9380:9100::/47 2400:9380:9121::/48 2400:9380:9201::/48 2400:9380:9202::/48 2400:9380:9220::/47 2400:9380:9240::/47 2400:9380:9250::/47 2400:9380:9260::/48 2400:9380:9271::/48 2400:9380:9272::/48 2400:9380:9280::/47 2400:9380:9282::/48 2400:9380:92b0::/45 2400:95e0::/48 2400:9600:8800::/48 2400:9620::/32 2400:a040::/32 2400:a860:1::/48 2400:a860:2::/47 2400:a860:4::/47 2400:a980::/29 2400:ae00:1981::/48 2400:b200::/32 2400:b600::/32 2400:b700::/48 2400:be00::/48 2400:cb80:e2e::/47 2400:cb80:e30::/44 2400:cb80:e40::/44 2400:da00::/32 2400:dd00::/28 2400:ebc0::/32 2400:ee00::/32 2400:f6e0::/32 2400:f720::/32 2400:fe00::/32 2401:1160::/32 2401:11a0:10::/44 2401:11a0:d150::/48 2401:11a0:d152::/48 2401:11a0:d158::/48 2401:1200::/48 2401:1320::/32 2401:1d40::/32 2401:20::/40 2401:2780::/32 2401:2e00::/32 2401:33c0::/32 2401:3480:2000::/48 2401:3480:3000::/36 2401:3480::/36 2401:34a0::/31 2401:3800::/32 2401:3880::/32 2401:3980::/32 2401:3a80::/32 2401:3b80::/32 2401:3c80::/32 2401:3d80::/32 2401:3e80::/32 2401:3f80::/32 2401:4180::/32 2401:4280::/32 2401:4380::/32 2401:4480::/32 2401:4580::/32 2401:4680::/32 2401:4780::/32 2401:4880::/32 2401:4a80::/32 2401:5c20:10::/48 2401:70e0::/32 2401:71c0::/48 2401:7660::/48 2401:7700::/32 2401:7d40::/32 2401:7e00::/32 2401:800::/32 2401:8be0::/48 2401:8d00:12::/48 2401:8d00:14::/48 2401:8d00:4::/48 2401:8d00:7::/48 2401:8d00:8::/47 2401:8d00::/46 2401:8d00:b::/48 2401:8d00:c::/48 2401:8d00:f::/48 2401:9a00:10::/46 2401:9a00::/44 2401:a140:1::/48 2401:b180::/32 2401:b400:11::/48 2401:b400:14::/48 2401:b400:16::/47 2401:b400:20::/47 2401:b400:8::/47 2401:b400::/45 2401:b680::/32 2401:bd60::/32 2401:be00::/32 2401:c020:14::/48 2401:c020:6::/48 2401:c020:8::/47 2401:ca00::/32 2401:cb80::/32 2401:cc00::/32 2401:ce00::/32 2401:d0e0:2::/47 2401:d180:10::/47 2401:d180:111::/48 2401:d180:2120::/48 2401:d180::/46 2401:d920::/48 2401:de00::/32 2401:ec00::/32 2401:f860:100::/40 2401:f860:86::/47 2401:f860:88::/47 2401:f860:90::/46 2401:f860:94::/47 2401:f860:a::/47 2401:f860:c::/46 2401:f860:f100::/40 2401:f860:f6::/48 2401:fa00:40::/43 2402:1440::/32 2402:2000::/32 2402:20e0:f000::/48 2402:3180:8000::/33 2402:3180::/46 2402:3c00::/32 2402:3f80:1400::/40 2402:4440::/32 2402:4b80::/32 2402:4e00::/32 2402:5e40::/32 2402:5ec0::/32 2402:6e80::/32 2402:6f40:2::/48 2402:6f40::/48 2402:6fc0::/48 2402:7040::/32 2402:7d80:240::/47 2402:7d80:6666::/48 2402:7d80:8888::/48 2402:7d80:9999::/48 2402:7d80::/48 2402:840:d000::/46 2402:840:e000::/46 2402:840:f000::/38 2402:8bc0::/32 2402:8cc0:200::/40 2402:8cc0::/40 2402:92c0::/48 2402:93c0:100::/48 2402:93c0:20::/48 2402:93c0::/48 2402:9a80::/32 2402:a200::/32 2402:b8c0:106::/48 2402:b8c0:186::/48 2402:b8c0:6::/48 2402:b8c0:86::/48 2402:b940:200::/39 2402:b940::/40 2402:c60::/48 2402:d340::/32 2402:db40:5100::/48 2402:db40:5900::/48 2402:db40:5f00::/46 2402:dfc0:50::/44 2402:dfc0::/44 2402:e380:100::/40 2402:e480::/32 2402:e740::/32 2402:e7c0::/32 2402:ef40::/32 2402:f000::/32 2402:f140:ff10::/46 2402:f140:ff14::/48 2402:f8c0:2::/48 2402:f8c0:5::/48 2402:f8c0::/47 2403:1b80::/48 2403:1ec0:1200::/48 2403:1ec0:1400::/48 2403:1ec0:1600::/48 2403:1ec0:1610::/48 2403:1ec0:1900::/48 2403:2040::/48 2403:2b40::/32 2403:3140::/32 2403:4240::/32 2403:4280::/47 2403:4300::/32 2403:4c80::/48 2403:5c80::/48 2403:600::/32 2403:6380:14::/47 2403:6380:40::/46 2403:6380:60::/44 2403:6a00::/32 2403:7580::/32 2403:8080:101::/48 2403:8c00::/32 2403:9b00::/32 2403:a100::/48 2403:a140:100::/40 2403:a140:10::/48 2403:a200::/32 2403:ac00::/32 2403:ad80:8008::/48 2403:b400::/32 2403:c80::/32 2403:c980::/32 2403:d400::/32 2403:f4c0::/48 2403:ffc0:1100::/40 2403:ffc0:1200::/39 2404:1c80::/32 2404:3700::/48 2404:4dc0::/32 2404:6380:1000::/48 2404:6380:8001::/48 2404:6380::/48 2404:6500:dcb3::/48 2404:7180:a000::/48 2404:7180:a010::/48 2404:7180:aa00::/48 2404:7180:aa10::/48 2404:7180:b001::/48 2404:7180:b002::/48 2404:7180:b010::/48 2404:7180:c001::/48 2404:7180:c002::/48 2404:7180:c011::/48 2404:7180:c012::/48 2404:7180:f000::/48 2404:7180:f010::/48 2404:7240::/33 2404:7600::/32 2404:7940::/32 2404:bc0:1::/48 2404:bc0:4000::/43 2404:bc0:4100::/43 2404:bc0:4200::/43 2404:bc0:4300::/44 2404:bc0:4400::/43 2404:bc0:4f00::/43 2404:c2c0:240::/44 2404:c2c0:280::/44 2404:c2c0:2c0::/44 2404:c2c0:4000::/40 2404:c2c0:4100::/48 2404:c2c0:501::/48 2404:c2c0:6000::/40 2404:c2c0:8000::/36 2404:c2c0::/40 2404:c2c0:bb00::/40 2404:c2c0:c000::/36 2404:c300::/32 2404:c940::/48 2404:e280::/47 2404:e5c0::/32 2404:e8c0::/32 2404:f4c0::/32 2405:1480:1000::/48 2405:1480:2000::/48 2405:1480:3000::/47 2405:3140:11::/48 2405:3140:21::/48 2405:3140:31::/48 2405:3140:3a::/48 2405:3bc0::/48 2405:66c0::/32 2405:68c0:21::/48 2405:6940::/48 2405:6c0:2::/48 2405:6c0:4::/48 2405:6f00:c101::/48 2405:6f00:c102::/48 2405:6f00:c170::/47 2405:6f00:c602::/48 2405:7040:6000::/47 2405:78c0:6e00::/43 2405:80:13::/48 2405:80:1::/48 2405:8280::/32 2405:8a40::/32 2405:a900:ffee::/48 2405:a900:fffe::/48 2405:ad00::/32 2405:b7c0::/32 2405:d900::/32 2405:e000::/32 2405:f580::/32 2405:f940::/32 2406:1080::/48 2406:1e40:f012::/47 2406:2700::/32 2406:280::/32 2406:3340::/32 2406:3640:1::/48 2406:3d80::/32 2406:4d00::/48 2406:52c0::/32 2406:5340:6666::/48 2406:5340:8888::/48 2406:5ac0::/32 2406:6100::/32 2406:840:8100::/40 2406:840:9000::/44 2406:840:9600::/43 2406:840:9620::/44 2406:840:9961::/48 2406:840:9962::/47 2406:840:996c::/48 2406:840:a7fe::/48 2406:840:a800::/37 2406:840:e080::/44 2406:840:e0cf::/48 2406:840:e0e0::/46 2406:840:e0e4::/47 2406:840:e0e8::/48 2406:840:e10f::/48 2406:840:e14f::/48 2406:840:e201::/48 2406:840:e230::/44 2406:840:e260::/48 2406:840:e2cf::/48 2406:840:e600::/45 2406:840:e608::/46 2406:840:e621::/48 2406:840:e666::/47 2406:840:e720::/44 2406:840:e770::/48 2406:840:e777::/48 2406:840:e80f::/48 2406:840:eb00::/46 2406:840:eb04::/47 2406:840:eb07::/48 2406:840:eb08::/48 2406:840:eb0b::/48 2406:840:eb0f::/48 2406:840:ee40::/47 2406:840:ee44::/48 2406:840:ee4b::/48 2406:840:ee4d::/48 2406:840:eee5::/48 2406:840:f200::/47 2406:840:f203::/48 2406:840:f380::/44 2406:840:f44f::/48 2406:840:fa01::/48 2406:840:fa40::/48 2406:840:fa60::/44 2406:840:fc10::/44 2406:840:fc20::/43 2406:840:fc40::/42 2406:840:fc80::/42 2406:840:fcc0::/44 2406:840:fcd0::/48 2406:840:fd40::/42 2406:840:fd80::/44 2406:840:fd9f::/48 2406:840:fda0::/43 2406:840:fdc0::/44 2406:840:fdd1::/48 2406:840:fde5::/48 2406:840:fde6::/47 2406:840:fe27::/48 2406:840:fe90::/46 2406:840:fe94::/48 2406:840:fe96::/47 2406:840:fe98::/46 2406:840:fe9d::/48 2406:840:fe9e::/48 2406:840:fea2::/47 2406:840:fea4::/46 2406:840:fea8::/45 2406:840:fec0::/48 2406:840:fec2::/47 2406:840:fec4::/47 2406:840:fec8::/48 2406:840:feca::/47 2406:840:fecc::/47 2406:840:fecf::/48 2406:840:fed1::/48 2406:840:fed8::/48 2406:840:fedb::/48 2406:840:fedc::/47 2406:840:fedf::/48 2406:840:fef0::/48 2406:840:fef3::/48 2406:840:fef4::/48 2406:840:fef6::/47 2406:840:fef8::/46 2406:840:fefc::/48 2406:840:feff::/48 2406:8880::/48 2406:b640:100::/48 2406:b640:4100::/48 2406:cac0:200::/40 2406:cac0::/40 2406:cf00:1000::/43 2406:cf00::/48 2406:d440:100::/44 2406:d440:200::/44 2406:d440:300::/44 2406:d440:ff00::/48 2406:d440:ffff::/48 2406:e3c0::/32 2406:e500::/33 2407:23c0::/48 2407:2840::/48 2407:3740::/48 2407:37c0::/32 2407:4980::/32 2407:5380::/32 2407:6c40:1100::/48 2407:6c40:1210::/48 2407:6c40:1500::/48 2407:6c40:1600::/40 2407:8f40:2::/48 2407:9f00::/32 2407:ad80::/32 2407:ae80::/32 2407:b380:8000::/48 2407:bc00::/32 2407:c080:4000::/37 2407:c080:5000::/37 2407:c080:6000::/36 2407:c080:8000::/36 2407:c080::/35 2407:d9c0::/32 2408:4000::/22 2408:8000:1000::/36 2408:8000:2000::/35 2408:8000:2::/47 2408:8000:4000::/34 2408:8000:8000::/33 2408:8000::/48 2408:8001::/32 2408:8020::/30 2408:8024::/31 2408:8026::/32 2408:802a:8000::/33 2408:802c::/32 2408:803e::/32 2408:8056::/32 2408:805c::/30 2408:8060::/33 2408:80c2::/33 2408:80c5::/33 2408:80ca:8000::/34 2408:80ca::/33 2408:80da:8000::/34 2408:80da::/33 2408:80e0:4000::/34 2408:80e0:8000::/33 2408:80e2::/33 2408:80e9:4000::/34 2408:80ea:4000::/34 2408:80ea:8000::/33 2408:80f0:4000::/34 2408:80f0:8000::/33 2408:80f1:100::/43 2408:80f1:120::/44 2408:80f1:160::/43 2408:80f1:180::/43 2408:80f1:1b0::/44 2408:80f1:1c0::/43 2408:80f1:1e0::/44 2408:80f1:200::/40 2408:80f1:40::/43 2408:80f1:70::/44 2408:80f1:80::/41 2408:80f1::/42 2408:80f5:4000::/34 2408:80f9:4000::/34 2408:80fa:4000::/34 2408:80fa:8000::/33 2408:8120:1::/48 2408:8120:2::/48 2408:8120:7000::/36 2408:8140:2000::/48 2408:815f:e000::/35 2408:8181:6000::/40 2408:8181:8000::/40 2408:8181:a000::/40 2408:8181:a220::/44 2408:8181:e000::/40 2408:8182:6000::/40 2408:8182:c000::/40 2408:8183:4000::/40 2408:8183:8000::/40 2408:81a2:2000::/35 2408:81a2:4000::/35 2408:81a3:6000::/35 2408:81a3:c800::/48 2408:81a3:ca66::/48 2408:8206::/31 2408:8208::/29 2408:8210::/30 2408:8214::/31 2408:821a::/31 2408:8220::/31 2408:8226::/32 2408:822a::/31 2408:822e::/31 2408:8230::/29 2408:8238::/31 2408:823c::/31 2408:8240::/32 2408:8244::/30 2408:8248::/30 2408:824c::/32 2408:824e::/31 2408:8250::/30 2408:8254::/32 2408:8256::/31 2408:8258::/30 2408:825c::/31 2408:825f::/32 2408:8260::/32 2408:8262::/31 2408:8264::/31 2408:8266::/32 2408:826a::/32 2408:826c::/30 2408:8270::/32 2408:8274::/30 2408:8278::/31 2408:827a::/32 2408:8306::/31 2408:8308::/30 2408:8310::/30 2408:832a::/32 2408:832e::/31 2408:8330::/30 2408:8338::/32 2408:8340::/32 2408:8344::/30 2408:8348::/30 2408:834e::/31 2408:8350::/30 2408:8354::/32 2408:8356::/31 2408:8358::/30 2408:8360::/30 2408:8364::/31 2408:836c::/30 2408:8374::/30 2408:8378::/31 2408:837a::/32 2408:8406:100::/41 2408:8406:1800::/42 2408:8406:180::/42 2408:8406:1840::/43 2408:8406:18c0::/42 2408:8406:1900::/41 2408:8406:1980::/42 2408:8406:2400::/42 2408:8406:2440::/43 2408:8406:24c0::/42 2408:8406:2500::/41 2408:8406:2580::/42 2408:8406:3000::/42 2408:8406:3040::/43 2408:8406:30c0::/42 2408:8406:3100::/41 2408:8406:3180::/42 2408:8406:3c00::/42 2408:8406:3c40::/43 2408:8406:3cc0::/42 2408:8406:3d00::/41 2408:8406:3d80::/42 2408:8406:40::/43 2408:8406:4800::/42 2408:8406:4840::/43 2408:8406:48c0::/42 2408:8406:4900::/41 2408:8406:4980::/42 2408:8406:5400::/42 2408:8406:5440::/43 2408:8406:54c0::/42 2408:8406:5500::/41 2408:8406:5580::/42 2408:8406:6000::/42 2408:8406:6040::/43 2408:8406:60c0::/42 2408:8406:6100::/41 2408:8406:6180::/42 2408:8406:6c00::/42 2408:8406:6c40::/43 2408:8406:6cc0::/42 2408:8406:6d00::/41 2408:8406:6d80::/42 2408:8406:7800::/42 2408:8406:7840::/43 2408:8406:78c0::/42 2408:8406:7900::/41 2408:8406:7980::/42 2408:8406:8400::/42 2408:8406:8440::/43 2408:8406:84c0::/42 2408:8406:8500::/41 2408:8406:8580::/42 2408:8406:9000::/42 2408:8406:9040::/43 2408:8406:90c0::/42 2408:8406:9100::/41 2408:8406:9180::/42 2408:8406:9c00::/42 2408:8406:9c40::/43 2408:8406:9cc0::/42 2408:8406:9d00::/41 2408:8406:9d80::/42 2408:8406::/42 2408:8406:a800::/42 2408:8406:a840::/43 2408:8406:a8c0::/42 2408:8406:a900::/41 2408:8406:a980::/42 2408:8406:b400::/42 2408:8406:b440::/43 2408:8406:b4c0::/42 2408:8406:b500::/41 2408:8406:b580::/42 2408:8406:c00::/42 2408:8406:c0::/42 2408:8406:c40::/43 2408:8406:cc0::/42 2408:8406:d00::/41 2408:8406:d80::/42 2408:8407:500::/43 2408:8409:100::/41 2408:8409:1800::/40 2408:8409:180::/42 2408:8409:1900::/41 2408:8409:1980::/42 2408:8409:19c0::/43 2408:8409:1c0::/43 2408:8409:2400::/40 2408:8409:2500::/41 2408:8409:2580::/42 2408:8409:25c0::/43 2408:8409:3000::/40 2408:8409:3100::/41 2408:8409:3180::/42 2408:8409:31c0::/43 2408:8409:3c00::/40 2408:8409:3d00::/41 2408:8409:3d80::/42 2408:8409:3dc0::/43 2408:8409:4800::/40 2408:8409:4900::/41 2408:8409:4980::/42 2408:8409:49c0::/43 2408:8409:5400::/40 2408:8409:5500::/41 2408:8409:5580::/42 2408:8409:55c0::/43 2408:8409:6000::/40 2408:8409:6100::/41 2408:8409:6180::/42 2408:8409:61c0::/43 2408:8409:6c00::/40 2408:8409:6d00::/41 2408:8409:6d80::/42 2408:8409:6dc0::/43 2408:8409:7800::/40 2408:8409:7900::/41 2408:8409:7980::/42 2408:8409:79c0::/43 2408:8409:8400::/40 2408:8409:8500::/41 2408:8409:8580::/42 2408:8409:85c0::/43 2408:8409:9000::/40 2408:8409:9100::/41 2408:8409:9180::/42 2408:8409:91c0::/43 2408:8409:9c00::/40 2408:8409:9d00::/41 2408:8409:9d80::/42 2408:8409:9dc0::/43 2408:8409::/40 2408:8409:a800::/40 2408:8409:a900::/41 2408:8409:a980::/42 2408:8409:a9c0::/43 2408:8409:b400::/40 2408:8409:b500::/41 2408:8409:b580::/42 2408:8409:b5c0::/43 2408:8409:c00::/40 2408:8409:d00::/41 2408:8409:d80::/42 2408:8409:dc0::/43 2408:840c:1100::/40 2408:840c:1a00::/40 2408:840c:1c00::/40 2408:840c:1e00::/40 2408:840c:200::/40 2408:840c:2700::/40 2408:840c:2900::/40 2408:840c:2b00::/40 2408:840c:3400::/40 2408:840c:3600::/40 2408:840c:3800::/40 2408:840c:400::/40 2408:840c:4e00::/40 2408:840c:5000::/40 2408:840c:5200::/40 2408:840c:5b00::/40 2408:840c:5d00::/40 2408:840c:5f00::/40 2408:840c:6800::/40 2408:840c:6a00::/40 2408:840c:6c00::/40 2408:840c:7500::/40 2408:840c:7700::/40 2408:840c:7900::/40 2408:840c:8200::/40 2408:840c:8400::/40 2408:840c:8600::/40 2408:840c:8f00::/40 2408:840c:9100::/40 2408:840c:9300::/40 2408:840c:9c00::/40 2408:840c:9e00::/40 2408:840c::/40 2408:840c:a000::/40 2408:840c:a900::/40 2408:840c:ab00::/40 2408:840c:ad00::/40 2408:840c:b600::/40 2408:840c:b800::/40 2408:840c:ba00::/40 2408:840c:c300::/40 2408:840c:c500::/40 2408:840c:c700::/40 2408:840c:d000::/40 2408:840c:d00::/40 2408:840c:d200::/40 2408:840c:d400::/40 2408:840c:dd00::/40 2408:840c:de00::/39 2408:840c:f00::/40 2408:840d:1100::/42 2408:840d:1300::/42 2408:840d:1a00::/42 2408:840d:1c00::/42 2408:840d:1e00::/42 2408:840d:2000::/42 2408:840d:200::/42 2408:840d:2700::/42 2408:840d:2900::/42 2408:840d:2b00::/42 2408:840d:2d00::/42 2408:840d:3400::/42 2408:840d:3600::/42 2408:840d:3800::/42 2408:840d:3a00::/42 2408:840d:400::/42 2408:840d:4e00::/42 2408:840d:5000::/42 2408:840d:5200::/42 2408:840d:5400::/42 2408:840d:5b00::/42 2408:840d:5d00::/42 2408:840d:5f00::/42 2408:840d:600::/42 2408:840d:6100::/42 2408:840d:6800::/42 2408:840d:6a00::/42 2408:840d:6c00::/42 2408:840d:6e00::/42 2408:840d:7500::/42 2408:840d:7700::/42 2408:840d:7900::/42 2408:840d:7b00::/42 2408:840d:8200::/42 2408:840d:8400::/42 2408:840d:8600::/42 2408:840d:8800::/42 2408:840d:8f00::/42 2408:840d:9100::/42 2408:840d:9300::/42 2408:840d:9500::/42 2408:840d:9c00::/42 2408:840d:9e00::/42 2408:840d::/42 2408:840d:a000::/42 2408:840d:a200::/42 2408:840d:a900::/42 2408:840d:ab00::/42 2408:840d:ad00::/42 2408:840d:af00::/42 2408:840d:b600::/42 2408:840d:b800::/42 2408:840d:ba00::/42 2408:840d:bc00::/42 2408:840d:c300::/42 2408:840d:c500::/42 2408:840d:c700::/42 2408:840d:c900::/42 2408:840d:d000::/42 2408:840d:d00::/42 2408:840d:d200::/42 2408:840d:d400::/42 2408:840d:d600::/42 2408:840d:dd00::/42 2408:840d:de00::/42 2408:840d:f00::/42 2408:840e:dd00::/40 2408:840e:de00::/39 2408:8410::/30 2408:8414::/31 2408:8417::/32 2408:8418::/32 2408:841a::/31 2408:841c::/31 2408:841e::/32 2408:8420::/31 2408:8422::/32 2408:8426::/31 2408:842a::/31 2408:842c::/32 2408:842e::/32 2408:8431::/32 2408:8434::/30 2408:8438::/31 2408:843c::/30 2408:8440::/31 2408:8444::/30 2408:8448::/32 2408:844b::/32 2408:844c::/30 2408:8452::/31 2408:8454::/32 2408:8456::/31 2408:8458::/30 2408:845c::/31 2408:8460::/30 2408:8464::/31 2408:8466::/32 2408:8469::/32 2408:846a::/31 2408:846c::/30 2408:8470::/31 2408:8474::/30 2408:8478::/31 2408:847a::/32 2408:84e1::/32 2408:84e2::/31 2408:84e4::/30 2408:84e9::/32 2408:84eb::/32 2408:84ec::/30 2408:84f0::/28 2408:856c::/31 2408:8606::/31 2408:8608::/29 2408:8610::/30 2408:8614::/31 2408:861a::/31 2408:861c::/32 2408:8620::/31 2408:8624::/31 2408:8626::/32 2408:862a::/31 2408:862d::/32 2408:862e::/31 2408:8630::/29 2408:8638::/31 2408:863c::/31 2408:8640::/32 2408:8642::/32 2408:8644::/30 2408:8648::/31 2408:864c::/32 2408:864e::/31 2408:8650::/30 2408:8656::/31 2408:8658::/30 2408:865c::/31 2408:865f::/32 2408:8660::/32 2408:8662::/31 2408:8664::/31 2408:8666::/32 2408:866a::/31 2408:866c::/30 2408:8670::/32 2408:8674::/30 2408:8678::/31 2408:867a::/32 2408:8706::/31 2408:8708::/29 2408:8710::/30 2408:8719::/32 2408:871a::/31 2408:8720::/30 2408:8726::/32 2408:872b::/32 2408:872f::/32 2408:8730::/30 2408:8734::/31 2408:8736::/32 2408:8738::/32 2408:873c::/31 2408:8740::/32 2408:8742::/32 2408:8744::/30 2408:8748::/29 2408:8752::/32 2408:8756::/31 2408:8758::/30 2408:875c::/32 2408:8760::/32 2408:8762::/31 2408:8764::/31 2408:8766::/32 2408:8768::/32 2408:876a::/32 2408:876c::/30 2408:8770::/32 2408:8772::/31 2408:8774::/32 2408:8776::/31 2408:8778::/31 2408:877a::/32 2408:877c::/30 2408:8806:40::/43 2408:8806::/42 2408:880c::/30 2408:8810::/30 2408:8814::/31 2408:8818::/31 2408:882c::/32 2408:883a::/32 2408:8844::/43 2408:8856::/31 2408:8858::/30 2408:8862::/31 2408:8864::/31 2408:8866::/32 2408:886e::/31 2408:8872::/32 2408:8878::/31 2408:887e::/32 2408:8906:20::/44 2408:8907:9000::/44 2408:890c::/31 2408:8912::/31 2408:8914::/30 2408:891c::/32 2408:8920::/32 2408:8924::/32 2408:892c::/32 2408:8936::/32 2408:893a::/32 2408:8940::/32 2408:8948::/32 2408:894c::/32 2408:894e::/32 2408:8956::/31 2408:8958::/30 2408:8962::/31 2408:8964::/31 2408:8966::/32 2408:896c::/32 2408:896e::/31 2408:8972::/32 2408:8978::/30 2408:897e::/32 2408:8a00:c000::/36 2408:8a00:d000::/37 2408:8a00:e000::/35 2408:8a01::/36 2408:8a02:b110::/44 2408:8a02:b120::/44 2408:8a04:8000::/36 2408:8a04:e000::/40 2408:8a05:6000::/35 2408:8a05:8000::/36 2408:8a06:100::/47 2408:8a06::/47 2408:8a21:4000::/35 2408:8a22:9200::/39 2408:8a22:9400::/38 2408:8a22:9800::/40 2408:8a22:9a00::/39 2408:8a22:9c00::/38 2408:8a22:a000::/37 2408:8a23:4000::/34 2408:8a24:4000::/34 2408:8a26:c000::/34 2408:8a27:4000::/35 2409:2000::/31 2409:2002::/32 2409:27fa::/48 2409:27fa:f000::/48 2409:27fb::/48 2409:27fc::/48 2409:27fe::/33 2409:8000::/20 240a:2000::/40 240a:2001:1000::/36 240a:2001:100::/40 240a:4002:1b00::/40 240a:4010:8000::/33 240a:4020:83a::/48 240a:4020:883a::/48 240a:4021:83a::/48 240a:4021:883a::/48 240a:4084:2000::/35 240a:4088:a000::/35 240a:408c:2000::/35 240a:4090:1000::/39 240a:4090:1200::/40 240a:4090:120::/48 240a:4090:2010::/48 240a:4090:2041::/48 240a:4090:2061::/48 240a:4090:250::/48 240a:4090:3000::/39 240a:4090:3200::/40 240a:4090:5000::/39 240a:4090:50::/48 240a:4090:5200::/40 240a:4090:7000::/39 240a:4090:7200::/40 240a:4093::/35 240a:4094:2000::/35 240a:409c:2000::/35 240a:40a4:2000::/35 240a:40ac:2000::/35 240a:40b0:283a::/48 240a:40b0:483a::/48 240a:40b0:683a::/48 240a:40b0:83a::/48 240a:40c0:8000::/43 240a:40c0:8200::/48 240a:40c0:8240::/48 240a:40c0:a000::/43 240a:40c0:c000::/43 240a:40c0:e000::/43 240a:40c1:2000::/43 240a:40c1:4000::/43 240a:40c1:6000::/43 240a:40c1:8000::/43 240a:40c1::/43 240a:40c1:a000::/43 240a:40c1:c000::/43 240a:40c1:e000::/43 240a:40c2:2000::/43 240a:40c2:4000::/43 240a:40c2:6000::/43 240a:40c2:8000::/43 240a:40c2::/43 240a:40c2:a000::/43 240a:40c2:c000::/43 240a:40c2:e000::/43 240a:40c3:2000::/43 240a:40c3:4000::/43 240a:40c3:6000::/43 240a:40c3:8000::/43 240a:40c3::/43 240a:40c3:c000::/43 240a:40c3:c200::/48 240a:40c3:c240::/48 240a:40c3:e000::/43 240a:40c4:2000::/43 240a:40c4:4000::/43 240a:40c4::/43 240a:4172::/31 240a:41b0::/34 240a:41b8:8800::/40 240a:41b8:a000::/40 240a:41b8:d400::/40 240a:41d6::/31 240a:41f2::/31 240a:420a::/31 240a:4224:9000::/44 240a:4224:a000::/44 240a:4224:d000::/44 240a:4224:e000::/44 240a:4230::/31 240a:4242::/31 240a:4280::/26 240a:42c0::/27 240a:42e0::/28 240a:42f0::/29 240a:42f8::/30 240a:6001::/48 240a:a000::/20 240a:c000::/20 240b:e001::/32 240b:e002::/31 240b:e004::/30 240b:e008::/29 240b:e010::/32 240c:4000::/22 240c:6:3::/48 240c:6::/47 240c:c000::/20 240c:f:1::/48 240d:4000::/21 240e::/20 2602:2e0:ff::/48 2602:f7ee:ee::/48 2602:f92a:a478::/48 2602:f92a:d1ff::/48 2602:f92a:dead::/48 2602:f92a:e100::/44 2602:f93b:400::/38 2602:f9ba:10c::/48 2602:f9ba:a8::/48 2602:feda:182::/47 2602:feda:1bf::/48 2602:feda:1d1::/48 2602:feda:1df::/48 2602:feda:2d0::/44 2602:feda:2f0::/44 2605:9d80:8001::/48 2605:9d80:8011::/48 2605:9d80:8021::/48 2605:9d80:8031::/48 2605:9d80:8041::/48 2605:9d80:8081::/48 2605:9d80:9003::/48 2605:9d80:9013::/48 2605:9d80:9023::/48 2605:9d80:9033::/48 2605:9d80:9042::/48 2605:9d80:9071::/48 2605:9d80:9092::/48 2620:57:4004::/48 2804:1e48:9001::/48 2804:1e48:9002::/48 2a01:f100:100::/48 2a01:f100:1f8::/48 2a01:ffc7:100::/40 2a03:5840:126::/48 2a04:3e00:1002::/48 2a04:f580:8010::/47 2a04:f580:8090::/48 2a04:f580:8210::/47 2a04:f580:8290::/48 2a04:f580:9010::/48 2a04:f580:9012::/47 2a04:f580:9020::/48 2a04:f580:9030::/48 2a04:f580:9040::/48 2a04:f580:9050::/48 2a04:f580:9060::/48 2a04:f580:9070::/48 2a04:f580:9080::/48 2a04:f580:9090::/48 2a04:f580:9210::/48 2a04:f580:9212::/47 2a04:f580:9220::/48 2a04:f580:9230::/48 2a04:f580:9240::/48 2a04:f580:9250::/48 2a04:f580:9260::/48 2a04:f580:9270::/48 2a04:f580:9280::/48 2a04:f580:9290::/48 2a05:1087::/32 2a05:dfc3:ff00::/40 2a06:1180:1000::/48 2a06:1281:8000::/36 2a06:3603::/32 2a06:3604::/30 2a06:9f81:4600::/43 2a06:9f81:4640::/44 2a06:9f81:4660::/44 2a06:a005:1c40::/44 2a06:a005:260::/43 2a06:a005:280::/43 2a06:a005:2a0::/44 2a06:a005:8d0::/44 2a06:a005:9e0::/44 2a06:a005:a13::/48 2a06:a005:e9a::/48 2a09:54c6:3000::/36 2a09:54c6:6000::/35 2a09:54c6:b000::/36 2a09:54c6:c000::/36 2a09:54c6:e000::/36 2a09:b280:ff81::/48 2a09:b280:ff83::/48 2a09:b280:ff84::/47 2a0a:2840::/30 2a0a:2845:aab8::/46 2a0a:2845:d647::/48 2a0a:2846::/48 2a0a:6040:ec00::/40 2a0a:6044:6600::/39 2a0a:6044:7a00::/40 2a0b:2542::/48 2a0b:4b81:1001::/48 2a0b:4e07:b8::/47 2a0c:9a40:84e0::/48 2a0c:b641:571::/48 2a0c:b641:d40::/44 2a0e:97c0:5ef::/48 2a0e:97c0:83f::/48 2a0e:9b00::/29 2a0e:aa01:1fff::/48 2a0e:aa06:440::/48 2a0e:aa06:490::/44 2a0e:aa06:4e0::/44 2a0e:aa06:500::/44 2a0e:aa06:520::/48 2a0e:aa06:525::/48 2a0e:aa06:541::/48 2a0e:aa06::/40 2a0e:aa07:e01b::/48 2a0e:aa07:e021::/48 2a0e:aa07:e025::/48 2a0e:aa07:e030::/48 2a0e:aa07:e035::/48 2a0e:aa07:e039::/48 2a0e:aa07:e044::/48 2a0e:aa07:e0e0::/44 2a0e:aa07:e151::/48 2a0e:aa07:e155::/48 2a0e:aa07:e160::/47 2a0e:aa07:e162::/48 2a0e:aa07:e16a::/48 2a0e:aa07:e1a0::/44 2a0e:aa07:e1e1::/48 2a0e:aa07:e1e2::/47 2a0e:aa07:e1e4::/47 2a0e:aa07:e1e6::/48 2a0e:aa07:e200::/44 2a0e:aa07:e210::/48 2a0e:aa07:e21c::/47 2a0e:aa07:e220::/44 2a0e:aa07:f0d0::/46 2a0e:aa07:f0d4::/47 2a0e:aa07:f0d8::/48 2a0e:aa07:f0de::/47 2a0e:b107:12b::/48 2a0e:b107:14a0::/44 2a0e:b107:178d::/48 2a0e:b107:178e::/48 2a0e:b107:272::/48 2a0e:b107:740::/44 2a0e:b107:c10::/48 2a0e:b107:da0::/44 2a0e:b107:dce::/48 2a0f:5707:ac00::/47 2a0f:7803:e300::/40 2a0f:7803:f5d0::/44 2a0f:7803:f5e0::/43 2a0f:7803:f680::/43 2a0f:7803:f6a0::/44 2a0f:7803:f7c0::/42 2a0f:7803:f800::/43 2a0f:7803:f840::/44 2a0f:7803:fa21::/48 2a0f:7803:fa22::/47 2a0f:7803:fa24::/46 2a0f:7803:faf3::/48 2a0f:7803:fe41::/48 2a0f:7803:fe44::/46 2a0f:7803:fe4e::/48 2a0f:7803:fe81::/48 2a0f:7803:fe82::/48 2a0f:7804:f650::/44 2a0f:7804:f9f0::/44 2a0f:7807::/32 2a0f:7d07::/32 2a0f:85c1:ba5::/48 2a0f:85c1:ca0::/44 2a0f:85c1:ce1::/48 2a0f:85c1:cf1::/48 2a0f:9400:6110::/48 2a0f:9400:7700::/48 2a0f:ac00::/29 2a0f:ea47:fc1d::/48 2a10:2f00:15a::/48 2a10:67c2:2::/48 2a10:ccc0:d00::/46 2a10:ccc0:d0a::/47 2a10:ccc0:d0c::/47 2a10:ccc6:66c4::/48 2a10:ccc6:66c6::/48 2a10:ccc6:66c8::/47 2a10:ccc6:66ca::/48 2a10:ccc6:66cc::/47 2a12:f8c3::/36 2a13:1800:10::/48 2a13:1800:300::/44 2a13:1800:80::/44 2a13:1800::/48 2a13:1801:180::/43 2a13:a5c3:ff10::/44 2a13:a5c3:ff21::/48 2a13:a5c3:ff50::/44 2a13:a5c7:1800::/40 2a13:a5c7:2100::/48 2a13:a5c7:2102::/48 2a13:a5c7:2121::/48 2a13:a5c7:2301::/48 2a13:a5c7:2302::/48 2a13:a5c7:23c0::/42 2a13:a5c7:2600::/40 2a13:a5c7:2801::/48 2a13:a5c7:2803::/48 2a13:a5c7:3108::/48 2a13:a5c7:31a0::/43 2a13:a5c7:3301::/48 2a13:a5c7:3304::/48 2a13:a5c7:3306::/47 2a13:aac4:f000::/44 2a14:4c41::/32 2a14:67c1:20::/44 2a14:67c1:702::/47 2a14:67c1:704::/48 2a14:67c1:70::/47 2a14:67c1:73::/48 2a14:67c1:74::/48 2a14:67c1:a010::/44 2a14:67c1:a020::/48 2a14:67c1:a023::/48 2a14:67c1:a024::/48 2a14:67c1:a02a::/48 2a14:67c1:a02f::/48 2a14:67c1:a040::/47 2a14:67c1:a061::/48 2a14:67c1:a064::/48 2a14:67c1:a090::/45 2a14:67c1:a099::/48 2a14:67c1:a100::/43 2a14:67c1:a125::/48 2a14:67c1:a144::/48 2a14:67c1:a150::/44 2a14:67c1:b000::/48 2a14:67c1:b065::/48 2a14:67c1:b066::/48 2a14:67c1:b068::/47 2a14:67c1:b100::/46 2a14:67c1:b105::/48 2a14:67c1:b107::/48 2a14:67c1:b130::/46 2a14:67c1:b134::/47 2a14:67c1:b140::/48 2a14:67c1:b4a1::/48 2a14:67c1:b4a2::/48 2a14:67c1:b4c0::/45 2a14:67c1:b4d0::/44 2a14:67c1:b4e0::/43 2a14:67c1:b500::/47 2a14:67c1:b549::/48 2a14:67c1:b561::/48 2a14:67c1:b563::/48 2a14:67c1:b566::/48 2a14:67c1:b581::/48 2a14:67c1:b582::/48 2a14:67c1:b588::/47 2a14:67c1:b590::/47 2a14:67c1:b599::/48 2a14:67c5:1900::/40 2a14:7580:740::/44 2a14:7580:750::/47 2a14:7580:9200::/40 2a14:7580:9400::/39 2a14:7580:9600::/46 2a14:7580:960c::/48 2a14:7580:d000::/37 2a14:7580:d800::/39 2a14:7580:da00::/40 2a14:7580:e200::/40 2a14:7580:fa01::/48 2a14:7580:fe00::/40 2a14:7580:fff4::/48 2a14:7580:fff7::/48 2a14:7580:fffa::/48 2a14:7581:3100::/40 2a14:7581:3400::/47 2a14:7581:9010::/44 2a14:7581:b10::/48 2a14:7581:b12::/48 2a14:7581:b15::/48 2a14:7581:b20::/46 2a14:7581:b32::/47 2a14:7581:b44::/48 2a14:7581:b46::/47 2a14:7581:b48::/48 2a14:7581:b4a::/48 2a14:7581:b60::/48 2a14:7581:b62::/47 2a14:7581:b64::/46 2a14:7581:b6c::/47 2a14:7581:b6e::/48 2a14:7581:b82::/47 2a14:7581:ba0::/48 2a14:7581:ba2::/47 2a14:7581:ba4::/48 2a14:7581:bbb::/48 2a14:7581:bc2::/48 2a14:7581:bcd::/48 2a14:7581:bff::/48 2a14:7581:ffb::/48 2a14:7581:ffd::/48 2a14:7583:f201::/48 2a14:7583:f203::/48 2a14:7583:f300::/46 2a14:7583:f304::/48 2a14:7583:f4fe::/48 2a14:7583:f500::/48 2a14:7583:f701::/48 2a14:7583:f702::/47 2a14:7583:f704::/47 2a14:7583:f707::/48 2a14:7583:f708::/48 2a14:7583:f743::/48 2a14:7583:f764::/48 2a14:7584::/36 2a14:7c0:4a01::/48 2c0f:f7a8:8011::/48 2c0f:f7a8:8050::/48 2c0f:f7a8:805f::/48 2c0f:f7a8:8150::/48 2c0f:f7a8:815f::/48 2c0f:f7a8:8211::/48 2c0f:f7a8:9010::/48 2c0f:f7a8:9020::/48 2c0f:f7a8:9041::/48 2c0f:f7a8:9210::/47 2c0f:f7a8:9220::/48
294coder/Efficient-MIF
5,496
utils/progress_utils.py
from typing import Callable, List, Union from rich.progress import Progress, TextColumn, BarColumn, TimeElapsedColumn, TimeRemainingColumn, SpinnerColumn from contextlib import nullcontext class EasyProgress: tbar: Progress = None task_desp_ids: dict[str, int] = {} @classmethod def console(cls): assert cls.tbar is not None, '`tbar` has not initialized' return cls.tbar.console @classmethod def close_all_tasks(cls): if cls.tbar is not None: for task_id in cls.tbar.task_ids: cls.tbar.stop_task(task_id) # set the task_id all unvisible cls.tbar.update(task_id, visible=False) @classmethod def easy_progress(cls, task_desciptions: list[str], task_total: list[int], tbar_kwargs: dict={}, task_kwargs: list[dict[str, Union[str, int]]]=None, is_main_process: bool=True, *, start_tbar: bool=True, debug: bool=False) -> tuple[Progress, Union[list[int], int]]: """get a rich progress bar Args: task_desciptions (list[str]): list of task descriptions of `len(task_desciptions)` tasks task_total (list[int]): list of length each task tbar_kwargs (dict, optional): kwargs for progress bar. Defaults to {}. task_kwargs (list[dict[str, Union[str, int]]], optional): task kwargs for each task. Defaults to None. is_main_process (bool, optional): if is main process. Defaults to True. start_tbar (bool, optional): start running progress bar when ini. Defaults to True. debug (bool, optional): debug mode, set progress bar to be unvisible. Defaults to False. Returns: tuple[Progress, Union[list[int], int]]: Progress bar and task ids """ def _add_task_ids(tbar: Progress, task_desciptions, task_total, task_kwargs): task_ids = [] if task_kwargs is None: task_kwargs = [{'visible': False}] * len(task_desciptions) for task_desciption, task_total, id_task_kwargs in zip(task_desciptions, task_total, task_kwargs): if task_desciption in list(EasyProgress.task_desp_ids.keys()): task_id = EasyProgress.task_desp_ids[task_desciption] task_ids.append(task_id) else: task_id = tbar.add_task(task_desciption, total=task_total, **id_task_kwargs) task_ids.append(task_id) EasyProgress.task_desp_ids[task_desciption] = task_id return task_ids if len(task_ids) > 1 else task_ids[0] def _new_tbar_with_task_ids(task_desciptions, task_total, task_kwargs): if is_main_process: if task_kwargs is not None: assert len(task_desciptions) == len(task_total) == len(task_kwargs) else: assert len(task_desciptions) == len(task_total) # if (console := tbar_kwargs.pop('console', None)) is not None: # console._color_system = console._detect_color_system() # if 'console' in tbar_kwargs: # tbar_kwargs['console']._color_system = tbar_kwargs['console']._detect_color_system() tbar = Progress(TextColumn("[progress.description]{task.description}"), BarColumn(), TextColumn("[progress.percentage]{task.percentage:>3.0f}%"), SpinnerColumn(), TimeRemainingColumn(), TimeElapsedColumn(), **tbar_kwargs) EasyProgress.tbar = tbar task_ids = _add_task_ids(tbar, task_desciptions, task_total, task_kwargs) return tbar, task_ids else: return nullcontext(), [None] * len(task_desciptions) if len(task_desciptions) > 1 else None def _cached_tbar_with_new_task_ids(task_desciptions, task_total, task_kwargs): if is_main_process: tbar = EasyProgress.tbar task_ids = [] if task_kwargs is None: task_kwargs = [{'visible': False}] * len(task_desciptions) task_ids = _add_task_ids(tbar, task_desciptions, task_total, task_kwargs) return tbar, task_ids else: return nullcontext(), [None] * len(task_desciptions) if len(task_desciptions) > 1 else None if not debug: if EasyProgress.tbar is not None: rets = _cached_tbar_with_new_task_ids(task_desciptions, task_total, task_kwargs) else: rets = _new_tbar_with_task_ids(task_desciptions, task_total, task_kwargs) if start_tbar and is_main_process and not EasyProgress.tbar.live._started: EasyProgress.tbar.start() return rets else: return nullcontext(), [None] * len(task_desciptions) if len(task_desciptions) > 1 else None
2977094657/DidaAPI
1,578
frontend/docs/api/pomodoros/focus-timeline.md
# 获取专注记录时间线 获取专注记录的时间线数据,支持分页获取历史记录。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/pomodoros/timeline` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 | 参数名 | 类型 | 必填 | 说明 | 示例 | |--------|------|------|------|------| | to | string | 否 | 分页参数:上一页最后一条记录的startTime,用于获取更早的数据 | 2025-04-22T08:43:31.000+0000 | ## 响应格式 ### 成功响应 ```json [ { "id": "string", "tasks": [ { "taskId": "string", "title": "string", "projectName": "string", "startTime": "2025-04-22T08:43:31.000+0000", "endTime": "2025-04-22T09:38:58.000+0000" } ], "startTime": "2025-04-22T08:43:31.000+0000", "endTime": "2025-04-22T09:38:58.000+0000", "pauseDuration": 0, "etag": "string", "type": 1, "added": false } ] ``` ### 响应字段说明 | 字段 | 类型 | 说明 | |------|------|------| | id | string | 专注记录ID | | tasks | array | 关联的任务列表 | | tasks[].taskId | string | 任务ID | | tasks[].title | string | 任务标题 | | tasks[].projectName | string | 项目名称 | | tasks[].startTime | string | 任务开始时间 | | tasks[].endTime | string | 任务结束时间 | | startTime | string | 专注开始时间 | | endTime | string | 专注结束时间 | | pauseDuration | number | 暂停时长(毫秒) | | etag | string | 版本标识 | | type | number | 专注类型(0=番茄专注,1=正计时专注) | | added | boolean | 是否已添加 | ## 分页说明 1. **首次请求**:不传 `to` 参数,获取最新的专注记录(约31条) 2. **获取更多**:使用上一页最后一条记录的 `startTime` 作为 `to` 参数 3. **时间格式**:`to` 参数使用ISO 8601格式,如 `2025-04-22T08:43:31.000+0000` 4. **自动转换**:系统会自动将时间转换为时间戳并调整时区
2977094657/DidaAPI
1,099
frontend/docs/api/pomodoros/focus-hour-distribution.md
# 获取专注时间按小时分布 获取指定日期范围内按小时分布的专注时间统计。 ## 接口信息 - **接口URL**: `https://api.dida365.com/api/v2/pomodoros/statistics/dist/clock/{start_date}/{end_date}` - **请求方法**: `GET` - **认证要求**: 需要登录认证 - **所属平台**: 滴答清单 ## 请求参数 | 参数名 | 类型 | 必填 | 说明 | 示例 | |--------|------|------|------|------| | start_date | string | 是 | 开始日期,格式: YYYYMMDD | 20250601 | | end_date | string | 是 | 结束日期,格式: YYYYMMDD | 20250630 | ## 响应格式 ### 成功响应 ```json { "0": 120, "1": 33, "15": 120, "16": 72, "17": 120, "18": 131, "19": 135, "20": 129, "21": 159, "22": 120, "23": 120 } ``` ### 响应字段说明 | 字段 | 类型 | 说明 | |------|------|------| | {hour} | number | 指定小时的专注时长(分钟),hour为24小时制(0-23) | ### 数据说明 - **时间格式**: 使用24小时制,0表示00:00-01:00,23表示23:00-24:00 - **时长单位**: 分钟 - **统计范围**: 指定日期范围内的总计专注时间 - **数据聚合**: 将整个时间段内相同小时的专注时间累加 ## 使用说明 1. 确保已完成认证获取会话 2. 提供有效的日期范围参数 3. 日期格式必须为 YYYYMMDD 4. 返回数据为对象格式,键为小时数(字符串),值为专注时长(数字) 5. 只有有专注记录的小时才会出现在响应中 6. 可用于生成24小时专注时间分布图表 ## 应用场景 - **时间管理分析**: 了解自己在一天中哪些时段最专注 - **工作习惯优化**: 根据专注时间分布调整工作安排 - **效率统计**: 分析不同时间段的专注效率 - **数据可视化**: 生成24小时专注热力图或柱状图
294coder/Efficient-MIF
16,764
utils/loss_utils.py
from functools import partial import random from typing import Sequence, Union, TYPE_CHECKING from einops import reduce from contextlib import contextmanager import kornia from kornia.filters import spatial_gradient import torch import torch.nn as nn import torch.nn.functional as F from torch import Tensor from torch.autograd import Variable import numpy as np from math import exp import sys sys.path.append('./') from utils.misc import is_main_process, exists, default from utils.log_utils import easy_logger if TYPE_CHECKING: from model.base_model import BaseModel logger = easy_logger() class PerceptualLoss(nn.Module): def __init__(self, percep_net="vgg", norm=True): super(PerceptualLoss, self).__init__() self.norm = norm self.lpips_loss = lpips.LPIPS(net=percep_net).cuda() def forward(self, x, y): # assert x.shape == y.shape loss = self.lpips_loss(x, y, normalize=self.norm) return torch.squeeze(loss).mean() def gaussian(window_size, sigma): gauss = torch.Tensor( [ exp(-((x - window_size // 2) ** 2) / float(2 * sigma**2)) for x in range(window_size) ] ) return gauss / gauss.sum() class MaxGradientLoss(torch.nn.Module): def __init__(self, mean_batch=True) -> None: super().__init__() self.register_buffer( "x_sobel_kernel", torch.FloatTensor([[-1, 0, 1], [-2, 0, 2], [-1, 0, 1]]).expand(1, 1, 3, 3), ) self.register_buffer( "y_sobel_kernel", torch.FloatTensor([[-1, -2, -1], [0, 0, 0], [1, 2, 1]]).expand(1, 1, 3, 3), ) self.mean_batch = mean_batch def forward(self, fuse, ir, vis): c = fuse.size(1) fuse_grad_x = F.conv2d(fuse, self.x_sobel_kernel, padding=1, groups=c) fuse_grad_y = F.conv2d(fuse, self.y_sobel_kernel, padding=1, groups=c) ir_grad_x = F.conv2d(ir, self.x_sobel_kernel, padding=1, groups=c) ir_grad_y = F.conv2d(ir, self.y_sobel_kernel, padding=1, groups=c) vis_grad_x = F.conv2d(vis, self.x_sobel_kernel, padding=1, groups=c) vis_grad_y = F.conv2d(vis, self.y_sobel_kernel, padding=1, groups=c) max_grad_x = torch.maximum(ir_grad_x, vis_grad_x) max_grad_y = torch.maximum(ir_grad_y, vis_grad_y) if self.mean_batch: max_gradient_loss = ( F.l1_loss(fuse_grad_x, max_grad_x) + F.l1_loss(fuse_grad_y, max_grad_y) ) / 2 else: x_loss_b = F.l1_loss(fuse_grad_x, max_grad_x, reduction="none").mean( dim=(1, 2, 3) ) y_loss_b = F.l1_loss(fuse_grad_y, max_grad_y, reduction="none").mean( dim=(1, 2, 3) ) max_gradient_loss = (x_loss_b + y_loss_b) / 2 return max_gradient_loss def create_window(window_size, channel, sigma=1.5): _1D_window = gaussian(window_size, sigma).unsqueeze(1) _2D_window = _1D_window.mm(_1D_window.t()).float().unsqueeze(0).unsqueeze(0) window = Variable( _2D_window.expand(channel, 1, window_size, window_size).contiguous() ) return window def mci_loss(pred, gt): return F.l1_loss(pred, gt.max(1, keepdim=True)[0]) def sf(f1, kernel_radius=5): """copy from https://github.com/tthinking/YDTR/blob/main/losses/__init__.py Args: f1 (torch.Tensor): image shape [b, c, h, w] kernel_radius (int, optional): kernel redius using calculate sf. Defaults to 5. Returns: loss: loss item. type torch.Tensor """ device = f1.device b, c, h, w = f1.shape r_shift_kernel = ( torch.FloatTensor([[0, 0, 0], [1, 0, 0], [0, 0, 0]]) .to(device) .reshape((1, 1, 3, 3)) .repeat(c, 1, 1, 1) ) b_shift_kernel = ( torch.FloatTensor([[0, 1, 0], [0, 0, 0], [0, 0, 0]]) .to(device) .reshape((1, 1, 3, 3)) .repeat(c, 1, 1, 1) ) f1_r_shift = F.conv2d(f1, r_shift_kernel, padding=1, groups=c) f1_b_shift = F.conv2d(f1, b_shift_kernel, padding=1, groups=c) f1_grad = torch.pow((f1_r_shift - f1), 2) + torch.pow((f1_b_shift - f1), 2) kernel_size = kernel_radius * 2 + 1 add_kernel = torch.ones((c, 1, kernel_size, kernel_size)).float().to(device) kernel_padding = kernel_size // 2 f1_sf = torch.sum( F.conv2d(f1_grad, add_kernel, padding=kernel_padding, groups=c), dim=1 ) return 1 - f1_sf def _ssim(img1, img2, window, window_size, channel, size_average=True): mu1 = F.conv2d(img1, window, padding=window_size // 2, groups=channel) mu2 = F.conv2d(img2, window, padding=window_size // 2, groups=channel) mu1_sq = mu1.pow(2) mu2_sq = mu2.pow(2) mu1_mu2 = mu1 * mu2 sigma1_sq = ( F.conv2d(img1 * img1, window, padding=window_size // 2, groups=channel) - mu1_sq ) sigma2_sq = ( F.conv2d(img2 * img2, window, padding=window_size // 2, groups=channel) - mu2_sq ) sigma12 = ( F.conv2d(img1 * img2, window, padding=window_size // 2, groups=channel) - mu1_mu2 ) C1 = 0.01**2 C2 = 0.03**2 ssim_map = ((2 * mu1_mu2 + C1) * (2 * sigma12 + C2)) / ( (mu1_sq + mu2_sq + C1) * (sigma1_sq + sigma2_sq + C2) ) if size_average: return ssim_map.mean() else: return ssim_map.mean(1).mean(1).mean(1) class HybridL1L2(torch.nn.Module): def __init__(self): super(HybridL1L2, self).__init__() self.l1 = torch.nn.L1Loss() self.l2 = torch.nn.MSELoss() self.loss = LossWarpper(l1=self.l1, l2=self.l2) def forward(self, pred, gt): loss, loss_dict = self.loss(pred, gt) return loss, loss_dict class HybridSSIMSF(torch.nn.Module): def __init__(self, channel, weighted_r=(1.0, 5e-2, 6e-4, 25e-5)) -> None: super().__init__() self.weighted_r = weighted_r def forward(self, fuse, gt): # fuse: [b, 1, h, w] vi = gt[:, 0:1] # [b, 1, h, w] ir = gt[:, 1:] # [b, 1, h, w] _ssim_f_ir = ssim_loss_ir(fuse, ir) _ssim_f_vi = ssim_loss_vi(fuse, vi) _sf_f_ir = sf_loss_ir(fuse, ir) _sf_f_vi = sf_loss_vi(fuse, vi) ssim_f_ir = self.weighted_r[0] * _ssim_f_ir ssim_f_vi = self.weighted_r[1] * _ssim_f_vi sf_f_ir = self.weighted_r[2] * _sf_f_ir sf_f_vi = self.weighted_r[3] * _sf_f_vi loss_dict = dict( ssim_f_ir=ssim_f_ir, ssim_f_vi=ssim_f_vi, sf_f_ir=sf_f_ir, sf_f_vi=sf_f_vi, ) loss = ssim_f_ir + ssim_f_vi + sf_f_ir + sf_f_vi return loss, loss_dict class HybridSSIMMCI(torch.nn.Module): def __init__(self, channel, weight_r=(1.0, 1.0, 1.0)) -> None: super().__init__() self.ssim = SSIMLoss(channel=channel) self.mci_loss = mci_loss self.weight_r = weight_r def forward(self, fuse, gt): # fuse: [b, 1, h, w] vi = gt[:, 0:1] # [b, 1, h, w] ir = gt[:, 1:] # [b, 1, h, w] _ssim_f_ir = self.weight_r[0] * self.ssim(fuse, ir) _ssim_f_vi = self.weight_r[1] * self.ssim(fuse, vi) _mci_loss = self.weight_r[2] * self.mci_loss(fuse, gt) loss = _ssim_f_ir + _ssim_f_vi + _mci_loss loss_dict = dict( ssim_f_ir=_ssim_f_ir, ssim_f_vi=_ssim_f_vi, mci_loss=_mci_loss, ) return loss, loss_dict def accum_loss_dict(ep_loss_dict: dict, loss_dict: dict): for k, v in loss_dict.items(): if k in ep_loss_dict: ep_loss_dict[k] += v else: ep_loss_dict[k] = v return ep_loss_dict def ave_ep_loss(ep_loss_dict: dict, ep_iters: int): for k, v in ep_loss_dict.items(): ep_loss_dict[k] = v / ep_iters return ep_loss_dict @is_main_process def ave_multi_rank_dict(rank_loss_dict: "list[dict] | dict"): # type is dict is only one process assert isinstance(rank_loss_dict, (list, dict)), 'rank_loss_dict must be a list or a dict' if isinstance(rank_loss_dict, dict): return rank_loss_dict n = len(rank_loss_dict) if n == 1: return rank_loss_dict[0] ave_dict = {} keys = rank_loss_dict[0].keys() for k in keys: vs = 0 for d in rank_loss_dict: v = d[k] vs = vs + v ave_dict[k] = vs / n return ave_dict class HybridL1SSIM(torch.nn.Module): def __init__(self, channel=31, weighted_r=(1.0, 0.1)): super(HybridL1SSIM, self).__init__() assert len(weighted_r) == 2 self._l1 = torch.nn.L1Loss() self._ssim = SSIMLoss(channel=channel) self.loss = LossWarpper(weighted_r, l1=self._l1, ssim=self._ssim) def forward(self, pred, gt): loss, loss_dict = self.loss(pred, gt) return loss, loss_dict class HybridCharbonnierSSIM(torch.nn.Module): def __init__(self, weighted_r, channel=31) -> None: super().__init__() self._ssim = SSIMLoss(channel=channel) self._charb = CharbonnierLoss(eps=1e-4) self.loss = LossWarpper(weighted_r, charbonnier=self._charb, ssim=self._ssim) def forward(self, pred, gt): loss, loss_dict = self.loss(pred, gt) return (loss,) class HybridMCGMCI(torch.nn.Module): def __init__(self, weight_r=(1.0, 1.0)) -> None: super().__init__() self.mcg = MaxGradientLoss() self.mci = mci_loss self.weight_r = weight_r def forward(self, pred, gt): vis = gt[:, 0:1] ir = gt[:, 1:] mcg_loss = self.mcg(pred, ir, vis) * self.weight_r[0] mci_loss = self.mci(pred, gt) * self.weight_r[1] loss_dict = dict(mcg=mcg_loss, mci=mci_loss) return mcg_loss + mci_loss, loss_dict def gradient(input): """ 求图像梯度, sobel算子 :param input: :return: """ filter1 = nn.Conv2d( kernel_size=3, in_channels=1, out_channels=1, bias=False, padding=1, stride=1 ) filter2 = nn.Conv2d( kernel_size=3, in_channels=1, out_channels=1, bias=False, padding=1, stride=1 ) filter1.weight.data = ( torch.tensor([[-1.0, 0.0, 1.0], [-2.0, 0.0, 2.0], [-1.0, 0.0, 1.0]]) .reshape(1, 1, 3, 3) .to(input.device) ) filter2.weight.data = ( torch.tensor([[1.0, 2.0, 1.0], [0.0, 0.0, 0.0], [-1.0, -2.0, -1.0]]) .reshape(1, 1, 3, 3) .to(input.device) ) g1 = filter1(input) g2 = filter2(input) image_gradient = torch.abs(g1) + torch.abs(g2) return image_gradient class LossWarpper(torch.nn.Module): def __init__(self, weighted_ratio=(1.0, 1.0), **losses): super(LossWarpper, self).__init__() self.names = [] assert len(weighted_ratio) == len(losses.keys()) self.weighted_ratio = weighted_ratio for k, v in losses.items(): self.names.append(k) setattr(self, k, v) def forward(self, pred, gt) -> tuple[torch.Tensor, dict[torch.Tensor]]: loss = 0.0 d_loss = {} for i, n in enumerate(self.names): l = getattr(self, n)(pred, gt) * self.weighted_ratio[i] loss += l d_loss[n] = l return loss, d_loss class TorchLossWrapper(torch.nn.Module): def __init__(self, weight_ratio: Union[tuple[float], list[float]], **loss) -> None: super().__init__() self.key = list(loss.keys()) self.loss = list(loss.values()) self.weight_ratio = weight_ratio assert len(weight_ratio) == len(loss.keys()) def forward(self, pred, gt): loss_total = 0.0 loss_d = {} for i, l in enumerate(self.loss): loss_i = l(pred, gt) * self.weight_ratio[i] loss_total = loss_total + loss_i k = self.key[i] loss_d[k] = loss_i return loss_total, loss_d class SSIMLoss(torch.nn.Module): def __init__( self, win_size=11, win_sigma=1.5, data_range=1, size_average=True, channel=3 ): super(SSIMLoss, self).__init__() self.window_size = win_size self.size_average = size_average self.channel = channel self.window = create_window(win_size, self.channel, win_sigma) self.win_sigma = win_sigma def forward(self, img1, img2): # print(img1.size()) (_, channel, _, _) = img1.size() if channel == self.channel and self.window.data.type() == img1.data.type(): window = self.window else: window = create_window(self.window_size, channel, self.win_sigma) if img1.is_cuda: window = window.cuda(img1.get_device()) window = window.type_as(img1) self.window = window self.channel = channel return 1 - _ssim( img1, img2, window, self.window_size, channel, self.size_average ) def ssim(img1, img2, win_size=11, data_range=1, size_average=True): (_, channel, _, _) = img1.size() window = create_window(win_size, channel) if img1.is_cuda: window = window.cuda(img1.get_device()) window = window.type_as(img1) return _ssim(img1, img2, window, win_size, channel, size_average) def elementwise_charbonnier_loss( input: Tensor, target: Tensor, eps: float = 1e-3 ) -> Tensor: """Apply element-wise weight and reduce loss between a pair of input and target. """ return torch.sqrt((input - target) ** 2 + (eps * eps)) class HybridL1L2(nn.Module): def __init__(self, cof=10.0): super(HybridL1L2, self).__init__() self.l1 = nn.L1Loss() self.l2 = nn.MSELoss() self.cof = cof def forward(self, pred, gt): return self.l1(pred, gt) / self.cof + self.l2(pred, gt) def get_loss(loss_type, channel=31, **kwargs: "Unpack[GetLossKwargsdType]"): if loss_type == "mse": criterion = nn.MSELoss() elif loss_type == "l1": criterion = TorchLossWrapper((1.,), l1=nn.L1Loss()) elif loss_type == "hybrid": criterion = HybridL1L2() elif loss_type == "smoothl1": criterion = nn.SmoothL1Loss() elif loss_type == "l1ssim": criterion = HybridL1SSIM(channel=channel, weighted_r=(1.0, 0.1)) else: raise NotImplementedError(f"loss {loss_type} is not implemented") return criterion if __name__ == "__main__": # loss = SSIMLoss(channel=31) # loss = CharbonnierLoss(eps=1e-3) # x = torch.randn(1, 31, 64, 64, requires_grad=True) # y = x + torch.randn(1, 31, 64, 64) / 10 # l = loss(x, y) # l.backward() # print(l) # print(x.grad) # import PIL.Image as Image # vi = ( # np.array( # Image.open( # "/media/office-401/Elements SE/cao/ZiHanCao/datasets/RoadScene_and_TNO/training_data/vi/FLIR_05857.jpg" # ).convert("L") # ) # / 255 # ) # ir = ( # np.array( # Image.open( # "/media/office-401/Elements SE/cao/ZiHanCao/datasets/RoadScene_and_TNO/training_data/ir/FLIR_05857.jpg" # ).convert("L") # ) # / 255 # ) # torch.cuda.set_device("cuda:0") # vi = torch.tensor(vi)[None, None].float() # .cuda() # ir = torch.tensor(ir)[None, None].float() # .cuda() # fuse = ((vi + ir) / 2).repeat_interleave(2, dim=0) # fuse.requires_grad_() # print(fuse.requires_grad) # gt = torch.cat((vi, ir), dim=1).repeat_interleave(2, dim=0) # fuse_loss = HybridSSIMRMIFuse(weight_ratio=(1.0, 1.0, 1.0), ssim_channel=1) torch.cuda.set_device("cuda:1") class FuseModel: def only_fusion_step(self, a, b): return a + b fuse_loss = DRMFFusionLoss(grad_op='sobel', reduce_label=True, pseudo_l1_const=0.).cuda() # fuse_loss = get_emma_fusion_loss(FuseModel()) # print(fuse_loss(fused, (vis, ir), mask)) # u2fusion_loss = U2FusionLoss().cuda() import time while True: fused = torch.randn(1, 3, 64, 64).cuda().requires_grad_() vis = torch.randn(1, 3, 64, 64).cuda().requires_grad_() ir = torch.randn(1, 3, 64, 64).cuda().requires_grad_() mask = torch.randint(0, 3, (1, 1, 64, 64)).cuda().float() loss = fuse_loss(fused, (vis, ir), mask=mask) loss[0].backward() print(loss) time.sleep(0.1) # fuse_loss = HybridPIALoss().cuda(1) # fuse_loss = CDDFusionLoss() # .cuda() # loss, loss_d = fuse_loss(fuse, gt) # loss.backward() # print(loss) # print(loss_d) # print(fuse.grad) # mcg_mci_loss = HybridMCGMCI() # print(mcg_mci_loss(fuse, gt))
281677160/openwrt-package
79,959
luci-app-passwall/root/usr/share/passwall/rules/gfwlist
000webhost.com 030buy.com 0rz.tw 1000giri.net 10beasts.net 10conditionsoflove.com 10musume.com 123rf.com 12bet.com 12vpn.com 12vpn.net 1337x.to 138.com 141hongkong.com 141jj.com 141tube.com 1688.com.au 173ng.com 177pic.info 17t17p.com 18board.com 18comic.org 18onlygirls.com 18p2p.com 18virginsex.com 1984bbs.com 1991way.com 1dumb.com 1e100.net 1eew.com 1lib.sk 1mobile.com 1point3acres.com 1pondo.tv 2-hand.info 2000fun.com 2008xianzhang.info 2021hkcharter.com 2047.name 2047.one 2049bbs.xyz 21andy.com 21sextury.com 228.net.tw 233abc.com 24hrs.ca 25u.com 2lipstube.com 2shared.com 2waky.com 3-a.net 30boxes.com 315lz.com 32red.com 36rain.com 3a5a.com 3arabtv.com 3boys2girls.com 3d-game.com 3proxy.ru 3ren.ca 3tui.net 404museum.com 466453.com 4bluestones.biz 4chan.com 4dq.com 4everproxy.com 4irc.com 4mydomain.com 4pu.com 4rbtv.com 4shared.com 4sqi.net 500px.com 500px.org 51.ca 51jav.org 51luoben.com 5278.cc 5299.tv 56cun04.jigsy.com 5i01.com 5isotoi5.org 5maodang.com 611study.com 611study.icu 63i.com 64museum.org 64tianwang.com 64wiki.com 66.ca 666kb.com 666pool.cn 69shuba.cx 6do.news 6do.world 6park.com 6parkbbs.com 6parker.com 6parknews.com 7capture.com 7cow.com 8-d.com 85cc.us 881903.com 888.com 888poker.com 89-64.org 89.64.charter.constitutionalism.solutions 8964museum.com 8news.com.tw 8z1.net 91dasai.com 91porn.com 91porny.com 91vps.club 92ccav.com 991.com 99btgc01.com 99cn.info 9bis.com 9bis.net 9cache.com 9gag.com 9news.com.au a-normal-day.com a248.e.akamai.net a5.com.ru aamacau.com abc.com abc.net.au abc.xyz abchinese.com abebooks.co.uk abebooks.com abematv.akamaized.net abitno.linpie.com ablwang.com aboluowang.com about.gitlab.com about.me abplive.com abs.edu acast.com accim.org accountkit.com aceros-de-hispania.com acevpn.com acg.rip acg18.me acgbox.org acgkj.com acgnx.se acmedia365.com acmetoy.com acnw.com.au actfortibet.org actimes.com.au activpn.com aculo.us adcex.com addictedtocoffee.de addons.mozilla.org addyoutube.com adelaidebbs.com admin.recaptcha.net admob.com adpl.org.hk ads-twitter.com adsense.com adult-sex-games.com adult.friendfinder.com adultfriendfinder.com advanscene.com advertfan.com advertisercommunity.com ae.hao123.com ae.org aei.org aenhancers.com aex.com af.mil afantibbs.com afr.com afreecatv.com agnesb.fr agoogleaday.com agro.hk ai.binwang.me aiosearch.com aiph.net airasia.com airconsole.com airitilibrary.com airvpn.org aisex.com aiss.anws.gov.tw ait.org.tw aiweiwei.com aiweiweiblog.com akademiye.org akamaihd.net akiba-online.com akiba-web.com akinator.com akow.org al-islam.com alabout.com alanhou.com alarab.qa alasbarricadas.org alforattv.net alhayat.com alicejapan.co.jp aliengu.com alive.bar aljazeera.com alkasir.com all4mom.org allcoin.com allconnected.co alldrawnsex.com allfinegirls.com allgirlmassage.com allgirlsallowed.org allgravure.com alliance.org.hk allinfa.com alljackpotscasino.com allmovie.com allowed.org almostmy.com alphaporno.com alternate-tools.com alternativeto.net altrec.com alvinalexander.com alwaysdata.com alwaysdata.net alwaysvpn.com am730.com.hk amazon.co.jp amazonvideo.com ameblo.jp americangreencard.com americanunfinished.com americorps.gov amiblockedornot.com amigobbs.net amitabhafoundation.us amnesty.org amnesty.org.hk amnesty.tw amnestyusa.org ampproject.org amtb-taipei.org amuletmc.com anchor.fm anchorfree.com ancsconf.org andfaraway.net android-x86.org android.com androidapksfree.com androidify.com androidtv.com andygod.com angela-merkel.de angelfire.com angola.org angularjs.org animecrazy.net aniscartujo.com annas-archive.org annas-archive.se annatam.com anobii.com anonfiles.com anontext.com anonymitynetwork.com anonymizer.com anonymouse.org anpopo.com answering-islam.org anthonycalzadilla.com anthropic.com antichristendom.com antiwave.net antpool.com anyporn.com anysex.com ao3.org aobo.com.au aofriend.com aojiao.org aomedia.org aomiwang.com apartmentratings.com apartments.com apat1989.org apetube.com api-secure.recaptcha.net api-verify.recaptcha.net api.ai api.pureapk.com api.recaptcha.net api.steampowered.com apiary.io apigee.com apk.support apkcombo.com apkmirror.com apkmonk.com apkplz.com apkpure.com apkpure.net app.box.com app.cloudcone.com app.smartmailcloud.com appadvice.com appbrain.com appdownloader.net appledaily.com appledaily.com.tw apps.evozi.com appshopper.com appsocks.net appspot.com appsto.re aptoide.com ar.hao123.com archive.fo archive.is archive.li archive.md archive.org archive.ph archive.today archive.vn archiveofourown.com archiveofourown.org archives.gov archives.gov.tw arctosia.com areca-backup.org arena.taipei arethusa.su arlingtoncemetery.mil art4tibet1998.org arte.tv artofpeacefoundation.org artstation.com artsy.net arvanstorage.ir asacp.org asdfg.jp asg.to asia-gaming.com asiaharvest.org asianage.com asianews.it asiansexdiary.com asiaone.com asiatgp.com ask.com askstudent.com askynz.net aspi.org.au aspistrategist.org.au assembla.com assets.bwbx.io assimp.org astrill.com atc.org.au atchinese.com atgfw.org athenaeizou.com atlanta168.com atnext.com auctions.yahoo.co.jp audacy.com auntology.fandom.com authorizeddns.net authorizeddns.org autodraw.com av-e-body.com av.com av.movie av01.tv avaaz.org avcool.com avdb.in avdb.tv avfantasy.com avg.com avgle.com avidemux.org avmo.pw avmoo.com avmoo.net avmoo.pw avoision.com avyahoo.com axios.com axureformac.com azerimix.com azirevpn.com azurewebsites.net b-ok.cc b.hatena.ne.jp b0ne.com babylonbee.com babynet.com.hk backchina.com backpackers.com.tw backtotiananmen.com bad.news badiucao.com badjojo.com badoo.com bahamut.com.tw baidu.jp baijie.org bailandaily.com baixing.me baizhi.org banana-vpn.com band.us bandcamp.com bandwagonhost.com bangbrosnetwork.com bangchen.net bangdream.space bangkokpost.com bangumi.moe bangyoulater.com bankmobilevibe.com bannedbook.org bannednews.org banorte.com baramangaonline.com barenakedislam.com barnabu.co.uk bartender.dowjones.com barton.de bastillepost.com bayvoice.net bb-chat.tv bbc.co.uk bbc.com bbc.in bbcchinese.com bbchat.tv bbci.co.uk bbg.gov bbkz.com bbnradio.org bbs-tw.com bbs.brockbbs.com bbs.cantonese.asia bbs.ecstart.com bbs.hanminzu.org bbs.huasing.org bbs.junglobal.net bbs.mikocon.com bbs.morbell.com bbs.mychat.to bbs.naixi.net bbs.nyinfor.com bbs.sina.com bbs.skykiwi.com bbs.sou-tong.org bbsdigest.com bbsland.com bbsmo.com bbsone.com bbtoystore.com bcc.com.tw bcchinese.net bcex.ca bcmorning.com bdsmvideos.net beaconevents.com bearteach.com bebo.com beeg.com beepool.com beepool.org beevpn.com behance.net behindkink.com beijing1989.com beijing2022.art beijingspring.com belamionline.com bell.wiki bemywife.cc beric.me berlinerbericht.de berlintwitterwall.com berm.co.nz bestgore.com bestpornstardb.com bestvpn.com bestvpnanalysis.com bestvpnforchina.net bestvpnserver.com bestvpnservice.com bestvpnusa.com bet365.com betaclouds.net betfair.com betterhash.net betternet.co bettervpn.com bettween.com betvictor.com bewww.net beyondfirewall.com bfnn.org bfsh.hk bgme.me bgvpn.com bianlei.com biantailajiao.com biblesforamerica.org bibox.com biedian.me big.one bigfools.com bigjapanesesex.com bigmoney.biz bignews.org bigone.com bigsound.org bild.de biliworld.com billypan.com binance.com binance.org binancezh.cc binux.me bird.so bit-z.com bit.do bit.ly bitbay.net bitchute.com bitcointalk.org bitcoinworld.com bitfinex.com bitget.com bithumb.com bitmex.com bitshare.com bitsnoop.com bitterwinter.org bitvise.com bitz.ai bizhat.com bjnewlife.org bjs.org bjzc.org bl-doujinsouko.com blacked.com blacklogic.com blackmagicdesign.com blackvpn.com blewpass.com blinkx.com blinw.com blip.tv blockcast.it blockcn.com blockedbyhk.com blockless.com blocktempo.com blog.cryptographyengineering.com blog.de blog.excite.co.jp blog.expofutures.com blog.fizzik.com blog.foolsmountain.com blog.fuckgfw233.org blog.goo.ne.jp blog.jackjia.com blog.jp blog.lester850.info blog.martinoei.com blog.pathtosharepoint.com blog.pentalogic.net blog.ranxiang.com blog.reimu.net blog.sogoo.org blog.soylent.com blog.syx86.com blog.taragana.com blog.tiney.com blog.youthwant.com.tw blogblog.com blogcatalog.com blogcity.me blogdns.org blogger.com blogimg.jp blogjav.net bloglines.com bloglovin.com blogspot.ae blogspot.al blogspot.am blogspot.ba blogspot.be blogspot.bg blogspot.ca blogspot.cat blogspot.ch blogspot.cl blogspot.co.uk blogspot.com blogspot.com.ar blogspot.com.au blogspot.com.br blogspot.com.by blogspot.com.co blogspot.com.cy blogspot.com.ee blogspot.com.eg blogspot.com.es blogspot.com.mt blogspot.com.ng blogspot.com.tr blogspot.com.uy blogspot.cz blogspot.de blogspot.dk blogspot.fi blogspot.fr blogspot.gr blogspot.hk blogspot.hr blogspot.hu blogspot.ie blogspot.in blogspot.is blogspot.it blogspot.jp blogspot.kr blogspot.li blogspot.lt blogspot.lu blogspot.md blogspot.mk blogspot.mx blogspot.my blogspot.nl blogspot.no blogspot.pe blogspot.pt blogspot.qa blogspot.ro blogspot.ru blogspot.se blogspot.sg blogspot.si blogspot.sk blogspot.sn blogspot.tw blogspot.ug blogtd.org bloodshed.net bloomberg.cn bloomberg.com bloomberg.de bloombergview.com bloomfortune.com blubrry.com blueangellive.com bmdru.com bnbstatic.com bnext.com.tw bnn.co bnrmetal.com boardreader.com bod.asia bodog88.com bolehvpn.net bolin.netfirms.com bonbonme.com bonfoundation.org bongacams.com boobstagram.com book.com.tw book.zi5.me bookdepository.com bookepub.com books.com.tw booktopia.com.au bookwalker.com.tw bootstrapcdn.com borgenmagazine.com bot.nu botanwang.com bowenpress.com boxpn.com boxun.com boxun.tv boxunclub.com boyangu.com boyfriendtv.com boysfood.com boysmaster.com br.hao123.com br.st brainyquote.com braumeister.org brave.com bravotube.net brazzers.com breached.to break.com breakgfw.com breaking911.com breakingtweets.com breakwall.net briian.com brill.com brizzly.com broadbook.com broadpressinc.com brookings.edu brucewang.net brutaltgp.com bsky.app bsky.network bsky.social bt4g.org bt4gprx.com bt95.com btaia.com btbit.net btbtav.com btbtt.co btbtt.me btc.com btc98.com btcbank.bank btctrade.im btdig.com btdigg.org btguard.com btku.me btku.org btspread.com btsynckeys.com budaedu.org buddhanet.com.tw buffered.com bullguard.com bullog.org bullogger.com bumingbai.net bunbunhk.com busayari.com business-humanrights.org business.page businessinsider.com businesstoday.com.tw businessweek.com busu.org busytrade.com buzzhand.com buzzhand.net buzzorange.com buzzsprout.com bvpn.com bwgyhw.com bwh1.net bx.in.th bybit.com bynet.co.il bypasscensorship.org byrut.org c-est-simple.com c-span.org c-spanvideo.org c.mi.com c100tibet.org c2cx.com c3pool.com cableav.tv cablegatesearch.net cachefly.com cachefly.net cachinese.com cacnw.com cactusvpn.com cafepress.com cahr.org.tw calameo.com calebelston.com calendarz.com calgarychinese.ca calgarychinese.com calgarychinese.net cam4.com cam4.jp cam4.sg camfrog.com campaign-archive.com campaignforuyghurs.org cams.com cams.org.sg canadameet.com canalporno.com canyu.org caobian.info caochangqing.com caoporn.us cap.org.hk carabinasypistolas.com cardinalkungfoundation.org carfax.com cari.com.my caribbeancom.com carmotorshow.com carousell.com.hk carrd.co cartoonmovement.com casadeltibetbcn.org casatibet.org.mx casino.williamhill.com casinobellini.com casinoking.com casinoriva.com castbox.fm catbox.moe catch22.net catchgod.com catholic.org.hk catholic.org.tw cathvoice.org.tw cato.org cattt.com caus.com cbc.ca cbs.ntu.edu.tw cbsnews.com cbtc.org.hk cccat.cc cccat.co ccfd.org.tw cchere.com ccim.org cclife.ca cclife.org cclifefl.org ccthere.com ccthere.net cctmweb.net cctongbao.com ccue.ca ccue.com ccvoice.ca ccw.org.tw cdbook.org cdef.org cdig.info cdjp.org cdn-images.mailchimp.com cdn-telegram.org cdn.arstechnica.net cdn.assets.lfpcontent.com cdn.helixstudios.net cdn.jwplayer.com cdn.printfriendly.com cdn.seatguru.com cdn.softlayer.net cdn.statically.io cdn1.lp.saboom.com cdnews.com.tw cdninstagram.com cdp1989.org cdp1998.org cdp2006.org cdpeu.org cdpuk.co.uk cdpweb.org cdpwu.org cdw.com cecc.gov cellulo.info cenews.eu centauro.com.br centerforhumanreprod.com centralnation.com centurys.net certificate-transparency.org certificate.revocationcheck.com cfhks.org.hk cfos.de cfr.org cftfc.com cgdepot.org cgst.edu change.org changeip.name changeip.net changeip.org changp.com channelnewsasia.com chanworld.org chaoex.com chaos.social character.ai chat.lmsys.org chatgpt.com chatnook.com chaturbate.com checkgfw.com chengmingmag.com chenguangcheng.com chenpokong.com chenpokongvip.com chenshan20042005.wordpress.com cherrysave.com chhongbi.org china-mmm.jp.net china-mmm.net china-review.com.ua china-week.com china.ucanews.com china101.com china18.org china21.com china21.org china5000.us chinaaffairs.org chinaaid.net chinaaid.org chinaaid.us chinachange.org chinachannel.hk chinademocrats.org chinadialogue.net chinadigitaltimes.net chinaelections.org chinafile.com chinafreepress.org chinagate.com chinagfw.org chinagonet.com chinahorizon.org chinahush.com chinainperspective.com chinalaborwatch.org chinalawandpolicy.com chinalawtranslate.com chinamule.com chinamz.org chinanewscenter.com chinapost.com.tw chinapress.com.my chinarightsia.org chinasmile.net chinasocialdemocraticparty.com chinasoul.org chinasucks.net chinatopsex.com chinatown.com.au chinauncensored.tv chinaview.wordpress.com chinaway.org chinaworker.info chinayouth.org.hk chinese-leaders.org chinese-memorial.org chinese.donga.com chinese.engadget.com chinese.irib.ir chinese.soifind.com chinesedaily.com chinesedailynews.com chinesedemocracy.com chinesegay.org chinesen.de chinesenews.net.au chinesepen.org chineseradioseattle.com chineseupress.com chingcheong.com chinman.net chithu.org chobit.cc chrdnet.com christianfreedom.org christianstudy.com christiantimes.org.hk christusrex.org chrlawyers.hk chrome.com chromecast.com chromeexperiments.com chromestatus.com chromium.org chuang-yen.org chubold.com chubun.com churchinhongkong.org chushigangdrug.ch ci-en.jp cici.com ciciai.com cienen.com cineastentreff.de cipfg.org cirosantilli.com citizencn.com citizenlab.ca citizenlab.org citizensradio.org city365.ca city9x.com citypopulation.de citytalk.tw civicparty.hk civilhrfront.org civiliangunner.com civilmedia.tw civitai.com cixiaoya.club ck101.com clarionproject.org classicalguitarblog.net claude.ai clb.org.hk cldr.unicode.org cleansite.biz cleansite.info cleansite.us clearharmony.net clearsurance.com clearwisdom.net clementine-player.org clinica-tibet.ru clipconverter.cc clipfish.de cloud.dify.ai cloud.mail.ru cloudflare-dns.com cloudflare-ipfs.com cloudfunctions.net club1069.com clubhouseapi.com clyp.it cmcn.org cmegroup.com cmi.org.tw cmp.hku.hk cms.gov cmule.com cmx.im cn-proxy.com cn.fmnnow.com cn.freeones.com cn.nytstyle.com cn.sandscotaicentral.com cn.shafaqna.com cn.streetvoice.com cn.theaustralian.com.au cn.uncyclopedia.wikia.com cn.uptodown.com cn6.eu cna.com.tw cnabc.com cnbbnews.wordpress.com cnbeta.com.tw cnd.org cnex.org.cn cnineu.com cnn.com cnnews.chosun.com cnpolitics.org cnproxy.com co.ng.mil coat.co.jp cobinhood.com cochina.org codeshare.io codeskulptor.org cofacts.tw coffeemanga.to coinbase.com coinbene.com coinex.com coingecko.com coingi.com coinmarketcap.com coinrail.co.kr cointiger.com cointobe.com coinut.com colacloud.net collateralmurder.com collateralmurder.org comefromchina.com comic-mega.me commandarms.com comments.app commentshk.com communistcrimes.org communitychoicecu.com comparitech.com compileheart.com compress.to connect.facebook.net conoha.jp contactmagazine.net contests.twilio.com convio.net cool18.com coolaler.com coolder.com coolloud.org.tw coolncute.com coolstuffinc.com copilot.microsoft.com corumcollege.com cos-moe.com cosplayjav.pl costco.com cotweet.com counter.social coursehero.com covenantswatch.org.tw coze.com cpj.org cpu-monkey.com cq99.us crackle.com crazypool.org crazys.cc crazyshit.com crbug.com crchina.org crd-net.org creaders.net creadersnet.com creativelab5.com cristyli.com crocotube.com crossfire.co.kr crossvpn.net crosswall.org croxyproxy.com crrev.com crucial.com crunchyroll.com cruxpool.com csdparty.com csis.org csmonitor.com css.pixnet.in csuchen.de csw.org.uk ct.org.tw ctao.org ctinews.com ctitv.com.tw ctowc.org cts.com.tw ctwant.com cuhkacs.org cuiweiping.net culture.tw cumlouder.com curvefish.com cusp.hk cutout.pro cutscenes.net cw.com.tw cyberghost.natado.com cyberghostvpn.com cynscribe.com d-fukyu.com d.cash d100.net d2bay.com d2pass.com dabr.co.uk dabr.eu dabr.me dabr.mobi dadazim.com dadi360.com dafabet.com dafagood.com dafahao.com dafoh.org daftporn.com dagelijksestandaard.nl daidostup.ru dailymail.co.uk dailymotion.com dailynews.sina.com dailysabah.com dailyview.tw dajiyuan.com dajiyuan.de dajiyuan.eu dalailama-archives.org dalailama.com dalailama.mn dalailama.ru dalailama80.org dalailamacenter.org dalailamafellows.org dalailamafilm.com dalailamafoundation.org dalailamahindi.com dalailamainaustralia.org dalailamajapanese.com dalailamaprotesters.info dalailamaquotes.org dalailamatrust.org dalailamavisit.org.nz dalailamaworld.com dalianmeng.org daliulian.org danbooru.donmai.us danke4china.net daodu14.jigsy.com daolan.net darktech.org darpa.mil darrenliuwei.com dashlane.com data-vocabulary.org data.gov.tw daum.net david-kilgour.com dawangidc.com daxa.cn daylife.com db.tt dcard.tw dcmilitary.com ddc.com.tw ddex.io ddns.info ddns.me.uk ddns.mobi ddns.ms ddns.name ddns.net ddns.us deadhouse.org deadline.com deaftone.com debug.com deck.ly deck.new decodet.co deepai.org deepmind.com deezer.com definebabe.com deja.com delcamp.net delicious.com demo.unlock-music.dev democrats.org demosisto.hk deno.dev depositphotos.com desc.se desipro.de dessci.com destiny.xfiles.to destroy-china.jp deutsche-welle.de developers.box.net deviantart.com deviantart.net devio.us devpn.com devv.ai dfn.org dharamsalanet.com dharmakara.net diaoyuislands.org difangwenge.org digiland.tw digisfera.com diigo.com dipity.com directcreative.com discoins.com disconnect.me discord.com discord.gg discordapp.com discordapp.net discuss.com.hk discuss4u.com dish.com disk.yandex.com disk.yandex.ru disneyplus.com disp.cc disqus.com dit-inc.us diyin.org dizhidizhi.com dizhuzhishang.com djangosnippets.org dl-laby.jp dl.box.net dlive.tv dlsite.com dlyoutube.com dm530.net dma.mil dmc.nico dmcdn.net dmhy.org dmm.co.jp dns-dns.com dns-stuff.com dns04.com dns05.com dns1.us dns2.us dns2go.com dnscrypt.org dnset.com dnsrd.com dnssec.net dnvod.tv doc.new docker.com docker.io docs.deno.com docs.new doctorvoice.org documentingreality.com dogfartnetwork.com dojin.com dolc.de dolf.org.hk domain.club.tw domaintoday.com.au dongtaiwang.com dongtaiwang.net dongyangjing.com dontfilter.us doom9.org doosho.com doourbest.org dorjeshugden.com dotplane.com dotsub.com dotvpn.com doub.io doubibackup.com doubiyunbackup.com doublethinklab.org douchi.space dougscripts.com doujincafe.com download.aircrack-ng.org download.cnet.com dphk.org dpool.top dpp.org.tw dpr.info dragonex.io dragonsprings.org dreamamateurs.com drepung.org drgan.net dropbooks.tv dropbox.com dropboxapi.com dropboxusercontent.com drtuber.com dscn.info dsmtp.com dstk.dk dtiblog.com dtic.mil dubox.com duck.com duckduckgo.com duckload.com duckmylife.com duga.jp duihua.org duihuahrjournal.org dumb1.com duping.net duplicati.com dupola.com dupola.net dushi.ca duyaoss.com dvdpac.com dvorak.org dw-world.com dw-world.de dw.com dw.de dweb.link dwnews.com dwnews.net dynamic-dns.net dynamicdns.biz dynamicdns.co.uk dynamicdns.me.uk dynamicdns.org.uk dynawebinc.com dyndns-ip.com dyndns-pics.com dyndns.org dyndns.pro dynssl.com dynu.com dynu.net dynupdate.no-ip.com dysfz.cc dzze.com e-classical.com.tw e-gold.com e-hentai.org e-hentaidb.com e-info.org.tw e-zone.com.hk e123.hk e621.net earlytibet.com earthcam.com earthvpn.com eastasiaforum.org easternlightning.org eastturkestan.com eastturkistan-gov.org eastturkistancc.org eastturkistangovernmentinexile.us easyca.ca easypic.com ebony-beauty.com ebook.hyread.com.tw ebookbrowse.com ebookee.com ebtcbank.com ecfa.org.tw echofon.com ecimg.tw ecministry.net economist.com edgecastcdn.net edicypages.com edmontonchina.cn edmontonservice.com edoors.com edubridge.com edupro.org edx-cdn.org eeas.europa.eu eesti.ee eevpn.com efcc.org.hk effers.com efksoft.com efukt.com eic-av.com eireinikotaerukai.com eisbb.com eksisozluk.com elconfidencial.com electionsmeter.com elgoog.im elpais.com eltondisney.com emaga.com emanna.com emilylau.org.hk empfil.com emule-ed2k.com emulefans.com emuparadise.me en.favotter.net en.hao123.com enanyang.my encrypt.me encyclopedia.com enewstree.com enfal.de englishforeveryone.org englishfromengland.co.uk englishpen.org enlighten.org.tw entermap.com epac.to episcopalchurch.org epochhk.com epochtimes-bg.com epochtimes-romania.com epochtimes.co.il epochtimes.co.kr epochtimes.com epochtimes.com.tw epochtimes.cz epochtimes.de epochtimes.fr epochtimes.it epochtimes.jp epochtimes.ru epochtimes.se epochtimestr.com epochweek.com epochweekly.com eporner.com equinenow.com erabaru.net eracom.com.tw eraysoft.com.tr erepublik.com erights.net eriversoft.com ernestmandel.org erodaizensyu.com erodoujinlog.com erodoujinworld.com eromanga-kingdom.com eromangadouzin.com eromon.net eroprofile.com eroticsaloon.net esg.t91y.com eslite.com esmtp.biz esurance.com etaa.org.au etadult.com etaiwannews.com etherdelta.com ethermine.org etherscan.com etherscan.io etizer.org etokki.com etools.ncol.com etowns.net etowns.org etsy.com ettoday.net etvonline.hk eucasino.com eulam.com eurekavpt.com euronews.com evchk.wikia.com everipedia.org evschool.net exam.gov.tw exblog.jp exchristian.hk exhentai.org exmo.com exmormon.org expatshield.com expecthim.com expekt.com experts-univers.com exploader.net expressvpn.com exrates.me extmatrix.com extremetube.com ey.gov.tw eyevio.jp eyny.com ezpeer.com ezua.com f-droid.org f2pool.com f8.com fa.gov.tw facebook.com facebook.de facebook.design facebook.hu facebook.in facebook.nl facebook.se facebookmail.com facebookquotes4u.com faceless.me facesofnyfw.com facesoftibetanselfimmolators.info factchecklab.org factpedia.org faith100.org faithfuleye.com faiththedog.info fakku.net fallenark.com falsefire.com falun-co.org falun-ny.net falunart.org falunasia.info falunau.org falunaz.net falundafa-dc.org falundafa-florida.org falundafa-nc.org falundafa-pa.net falundafa.org falundafaindia.org falundafamuseum.org falungong.club falungong.de falungong.org.uk falunhr.org faluninfo.de faluninfo.net falunpilipinas.net familyfed.org famunion.com fan-qiang.com fanbox.cc fangeming.com fangeqiang.com fanglizhi.info fangong.forums-free.com fangong.org fangongheike.com fanhaodang.com fanhaolou.com fanqiang.network fanqiang.tk fanqiangdang.com fanqianghou.com fanqiangzhe.com fanswong.com fantv.hk fanyue.info fapdu.com faproxy.com faqserv.com fartit.com farwestchina.com fast.com fast.wistia.com fastestvpn.com fastpic.ru fastssh.com faststone.org fatbtc.com favstar.fm fawanghuihui.org faydao.com faz.net fb.com fb.me fb.watch fbaddins.com fbcdn.net fbsbx.com fbworkmail.com fc2.com fc2blog.net fc2china.com fc2cn.com fda.gov.tw fdc64.de fdc64.org fdc89.jp feedburner.com feeder.co feedly.com feeds.fileforum.com feedx.net feelssh.com feer.com feitian-california.org feitianacademy.org feixiaohao.com feministteacher.com fengzhenghu.com fengzhenghu.net fevernet.com ff.im fffff.at fflick.com ffvpn.com fgmtv.net fgmtv.org fhreports.net fiddle.jshell.net figprayer.com fileflyer.com fileserve.com filesor.com fillthesquare.org filmingfortibet.org filthdump.com financetwitter.com financialexpress.com finchvpn.com findmespot.com findyoutube.com findyoutube.net fingerdaily.com firearmsworld.net firebaseio.com fireofliberty.info fireofliberty.org firetweet.io firstpost.com firstrade.com fish.audio flagsonline.it flecheinthepeche.fr fleshbot.com fleursdeslettres.com flexpool.io flgjustice.org flickr.com flickrhivemind.net flickriver.com fling.com flipboard.com flipkart.com flitto.com flnet.org flog.tw flowhongkong.net flypool.org flyvpn.com flyzy2005.com fnac.be fnac.com fnc.ebc.net.tw fochk.org focustaiwan.tw focusvpn.com fofg.org fooooo.com forbes.com foreignaffairs.com foreignpolicy.com form.new forms.new forum.baby-kingdom.com forum.cyberctm.com forum.mymaji.com forum.palmislife.com forum.slime.com.tw forum.tvb.com forum.xinbao.de forum4hk.com fountmedia.io fourthinternational.org foxgay.com foxsub.com foxtang.com fpmt-osel.org fpmt.org fpmt.tw fpmtmexico.org fq.wikia.com fqrouter.com frank2019.me franklc.com freakshare.com free-gate.org free-hada-now.org free-proxy.cz free-ss.site free.bg free.com.tw free.fr freebeacon.com freebrowser.org freechal.com freechina.net freechina.news freeddns.com freeddns.org freedomcollection.org freedomhouse.org freedominfonetweb.wordpress.com freedomsherald.org freeforums.org freegao.com freehongkong.org freeilhamtohti.org freekazakhs.org freelotto.com freeman2.com freemoren.com freemorenews.com freemuse.org freenet-china.org freenetproject.org freenewscn.com freeopenvpn.com freeoz.org freess.org freessh.us freetcp.com freetibet.net freetibet.org freetibetanheroes.org freetls.fastly.net freetribe.me freeviewmovies.com freevpn.me freevpn.nl freewallpaper4.me freewebs.com freewechat.com freewww.info freexinwen.com freeyellow.com freezhihu.org friendfeed.com friends-of-tibet.org friendsoftibet.org fril.jp fring.com fringenetwork.com from-pr.com from-sd.com fromchinatousa.net frommel.net frontlinedefenders.org frootvpn.com froth.zone fscked.org fsurf.com ft.com ftchinese.com ftp1.biz ftpserver.biz ftv.com.tw ftvnews.com.tw ftx.com fucd.com fuchsia.dev fuckccp.com fuckccp.xyz fuckgfw.org fulione.com fullerconsideration.com fullservicegame.com funf.tw funkyimg.com funp.com fuq.com furbo.org furhhdl.org furinkan.com furrybar.com futurechinaforum.org futuremessage.org fux.com fuyindiantai.org fuyu.org.tw fw.cm fxcm-chinese.com fxnetworks.com g-area.org g-queen.com g.co g0v.social g6hentai.com gab.com gabocorp.com gaeproxy.com gaforum.org gagaoolala.com galaxymacau.com galenwu.com galstars.net game735.com gamebase.com.tw gamejolt.com gamer.com.tw gamez.com.tw gamousa.com ganges.com ganjing.com ganjingworld.com gaoming.net gaopi.net gardennetworks.com gardennetworks.org gartlive.com garudalinux.org gate.io gatecoin.com gather.com gatherproxy.com gaybubble.com gaycn.net gayhub.com gaymap.cc gaymenring.com gaytube.com gaywatch.com gazotube.com gcc.org.hk gclubs.com gcmasia.com gcpnews.com gcr.io gdaily.org gdzf.org geek-art.net geekerhome.com gekikame.com gelbooru.com generated.photos genius.com geocities.co.jp geocities.com geocities.jp geph.io gerefoundation.org get.app get.dev get.how get.page getastrill.com getchu.com getcloak.com getfoxyproxy.org getgom.com geti2p.net getiton.com getlantern.org getmalus.com getmdl.io getoutline.org getsession.org getsocialscope.com getsync.com gettr.com gettrials.com getuploader.com gfbv.de gfsale.com gfw.press gfw.report gfwatch.org ggpht.com ggssl.com ghidra-sre.org ghostpath.com ghut.org giantessnight.com gifree.com giga-web.jp gigporno.ru girlbanker.com git.io gitbook.io github.blog github.com github.io githubassets.com githubcopilot.com githubusercontent.com gitlab.net gizlen.net gjczz.com glarity.app glass8.eu global.ssl.fastly.net globaljihad.net globalmediaoutreach.com globalmuseumoncommunism.org globalrescue.net globaltm.org globalvoices.org globalvoicesonline.org globalvpn.net glock.com gloryhole.com glorystar.me gluckman.com glype.com gmail.com gmgard.com gmll.org gmodules.com gmp4.com gnci.org.hk gnews.org go-to-zlibrary.se go141.com go5.dev goagent.biz godaddy.com godfootsteps.org godoc.org godsdirectcontact.co.uk godsdirectcontact.org godsdirectcontact.org.tw godsimmediatecontact.com gofundme.com gohappy.com.tw gojet.krtco.com.tw gokbayrak.com golang.org goldbet.com goldbetsports.com golden-ages.org goldeneyevault.com goldenfrog.com goldstep.net goldwave.com gongm.in goo.gl goo.gle goo.ne.jp good.news gooday.xyz goodhope.school goodnewsnetwork.org goodreaders.com goodreads.com goodtv.com.tw goodtv.tv goofind.com google.ad google.ae google.al google.am google.as google.at google.az google.ba google.be google.bf google.bg google.bi google.bj google.bs google.bt google.by google.ca google.cat google.cd google.cf google.cg google.ch google.ci google.cl google.cm google.cn google.co.ao google.co.bw google.co.ck google.co.cr google.co.id google.co.il google.co.in google.co.jp google.co.ke google.co.kr google.co.ls google.co.ma google.co.mz google.co.nz google.co.th google.co.tz google.co.ug google.co.uk google.co.uz google.co.ve google.co.vi google.co.za google.co.zm google.co.zw google.com google.com.af google.com.ag google.com.ai google.com.ar google.com.au google.com.bd google.com.bh google.com.bn google.com.bo google.com.br google.com.bz google.com.co google.com.cu google.com.cy google.com.do google.com.ec google.com.eg google.com.et google.com.fj google.com.gh google.com.gi google.com.gt google.com.hk google.com.jm google.com.kh google.com.kw google.com.lb google.com.ly google.com.mm google.com.mt google.com.mx google.com.my google.com.na google.com.nf google.com.ng google.com.ni google.com.np google.com.om google.com.pa google.com.pe google.com.pg google.com.ph google.com.pk google.com.pr google.com.py google.com.qa google.com.sa google.com.sb google.com.sg google.com.sl google.com.sv google.com.tj google.com.tr google.com.tw google.com.ua google.com.uy google.com.vc google.com.vn google.cv google.cz google.de google.dev google.dj google.dk google.dm google.dz google.ee google.es google.fi google.fm google.fr google.ga google.ge google.gg google.gl google.gm google.gp google.gr google.gy google.hn google.hr google.ht google.hu google.ie google.im google.iq google.is google.it google.je google.jo google.kg google.ki google.kz google.la google.li google.lk google.lt google.lu google.lv google.md google.me google.mg google.mk google.ml google.mn google.ms google.mu google.mv google.mw google.ne google.nl google.no google.nr google.nu google.pl google.pn google.ps google.pt google.ro google.rs google.ru google.rw google.sc google.se google.sh google.si google.sk google.sm google.sn google.so google.sr google.st google.td google.tg google.tk google.tl google.tm google.tn google.to google.tt google.vg google.vu google.ws googleapis.com googleapps.com googlearth.com googleartproject.com googleblog.com googlebot.com googlechinawebmaster.com googlecode.com googlecommerce.com googledomains.com googledrive.com googleearth.com googlefiber.net googlegroups.com googlehosted.com googleideas.com googleinsidesearch.com googlemail.com googlemashups.com googlepagecreator.com googleplay.com googleplus.com googlescholar.com googlesource.com googleusercontent.com googlevideo.com googleweblight.com googlezip.net gopetition.com goreforum.com goregrish.com gospelherald.com got-game.org gotdns.ch gotgeeks.com gotquestions.org gotrusted.com gotw.ca gov.ir gov.taipei gov.tw gr8domain.biz gr8name.biz grammaly.com grandtrial.org graph.org graphis.ne.jp graphql.org gravatar.com greasyfork.org greatfire.org greatfire.us7.list-manage.com greatfirewall.biz greatfirewallofchina.org greatroc.org greatzhonghua.org greenparty.org.tw greenpeace.com.tw greenpeace.org greenreadings.com greenvpn.net greenvpn.org grindr.com grok.com ground.news gs-discuss.com gsearch.media gsp.target.com gstatic.com gtricks.com gtv.org gtv1.org gu-chu-sum.org guaguass.com guancha.org guangming.com.my guardster.com guishan.org gumroad.com gun-world.net gunsamerica.com gunsandammo.com guruonline.hk gutteruncensored.com gvlib.com gvm.com.tw gvt1.com gvt3.com gwins.org gwtproject.org gyalwarinpoche.com gyatsostudio.com gzm.tv gzone-anime.info h-china.org h-moe.com h1n1china.org h528.com h5dm.com h5galgame.me hacken.cc hacker.org hackmd.io hackthatphone.net hahlo.com haijiao.com haiwaikan.com hakkatv.org.tw halktv.com.tr handcraftedsoftware.org hanime.tv hanime1.me hao.news haproxy.org hardsextube.com hautelook.com hautelookcdn.com have8.com hbg.com hbo.com hclips.com hd.stheadline.com hdtvb.net hdzog.com heartyit.com heavy-r.com hec.su hecaitou.net hechaji.com heeact.edu.tw hegre-art.com helloandroid.com helloqueer.com hellouk.org helpeachpeople.com helpster.de helpzhuling.org hentai.to hentaitube.tv hentaivideoworld.com heqinglian.net heritage.org herokuapp.com herominers.com hexieshe.com hexieshe.xyz hexxeh.net heyuedi.com heyzo.com hgseav.com hhdcb3office.org hhthesakyatrizin.org hi-on.org.tw hiccears.com hidden-advent.org hide.me hidecloud.com hideipvpn.com hideman.net hideme.nl hidemy.name hidemyass.com hidemycomp.com higfw.com highpeakspureearth.com highrockmedia.com hiitch.com hikinggfw.org hilive.tv himalayan-foundation.org himalayanglacier.com himemix.com hindustantimes.com hinet.net hitbtc.com hitomi.la hiveon.net hiwifi.com hizb-ut-tahrir.info hizb-ut-tahrir.org hizbuttahrir.org hjclub.info hk-pub.com hk.frienddy.com hk.geocities.com hk.gradconnection.com hk.hao123img.com hk.jiepang.com hk01.com hka8964.wordpress.com hkacg.com hkacg.net hkanews.wordpress.com hkatvnews.com hkbc.net hkbf.org hkbookcity.com hkchronicles.com hkchurch.org hkci.org.hk hkcmi.edu hkcnews.com hkcoc.com hkcoc.weather.com.hk hkdailynews.com.hk hkday.net hkdc.us hkdf.org hkej.com hkepc.com hket.com hkfaa.com hkfront.org hkgalden.com hkgolden.com hkgpao.com hkheadline.com hkhkhk.com hkhrc.org.hk hkjc.com hkjp.org hklft.com hklts.org.hk hkmap.live hkopentv.com hkpeanut.com hkptu.org hkreporter.com hkreporter.loved.hk hmoegirl.com hmv.co.jp hmvdigital.ca hmvdigital.com hnjhj.com hnntube.com hojemacau.com.mo hola.com hola.org hole.thu.monster holyspiritspeaks.org home.saxo home.sina.com homedepot.com homeperversion.com homeservershow.com hongkong.fandom.com hongkongfp.com hongmeimei.com hongzhi.li honven.xyz hootsuite.com hoover.org hoovers.com hopto.org hornygamer.com hornytrip.com horrorporn.com hostloc.com hotair.com hotav.tv hotcoin.com hotels.cn hotfrog.com.tw hotgoo.com hotpot.hk hotshame.com hotspotshield.com hottg.com hotvpn.com howtoforge.com hoxx.com hoy.tv hpa.gov.tw hpjav.com hqcdp.org hqjapanesesex.com hqmovies.com hqsbnet.wordpress.com hqsbonline.wordpress.com hrcchina.org hrea.org hrichina.org hrntt.org hrtsea.com hrw.org hrweb.org hsex.men hsjp.net hsselite.com hst.net.tw hstern.net hstt.net ht.ly htkou.net htl.li html5rocks.com https443.net https443.org huaglad.com huanghuagang.org huangyiyu.com huaren.us huaren4us.com huashangnews.com huaxiabao.org huaxin.ph huayuworld.org hudatoriq.web.id hudson.org huffingtonpost.com huffpost.com huggingface.co hugoroy.eu huhaitai.com huhamhire.com huhangfei.com hulkshare.com hulu.com huluim.com humanparty.me humanrightspressawards.org hung-ya.com huobi.co huobi.com huobi.me huobi.pro huobi.sc huobipool.com huobipro.com huping.net hurgokbayrak.com hurriyet.com.tr hustlercash.com hut2.ru hutianyi.net hutong9.net huyandex.com hwadzan.tw hwayue.org.tw hxwk.org hxwq.org hybrid-analysis.com hyperrate.com hypothes.is i-cable.com i-part.com.tw i-scmp.com i.111666.best i.lithium.com i2p2.de i818hk.com iam.soy iamtopone.com iask.ca iav19.com iavian.net ibiblio.org ibros.org ibvpn.com icams.com ice.audionow.com icedrive.net icij.org icl-fi.org icoco.com iconpaper.org icu-project.org id.hao123.com id.heroku.com idemocracy.asia identi.ca idiomconnection.com idope.se idouga.com idv.tw ied2k.net ienergy1.com ifan.cz.cc ifcss.org ifjc.org ifreechina.wordpress.com ifreewares.com ift.tt igcd.net igfw.net igmg.de igoogle.com igotmail.com.tw igvita.com ihao.org iicns.com iipdigital.usembassy.gov ikstar.com ikwb.com ilbe.com ilhamtohtiinstitute.org illawarramercury.com.au illusionfactory.com ilove80.be ilovelongtoes.com im88.tw imageab.com imagefap.com imageflea.com imageglass.org images-gaytube.com imageshack.us imagevenue.com imagezilla.net imago-images.com imb.org imdb.com img.dlsite.jp img.ly img.picgo.net imgasd.com imgchili.net imgmega.com imgur.com imkev.com imlive.co imlive.com immigration.gov.tw immoral.jp impact.org.au improd.works in-disguise.com in99.org incapdns.net incloak.com incredibox.fr independent.co.uk india.com indiablooms.com indianarrative.com indiandefensenews.in indiatoday.in indiemerch.com inews-api.tvb.com info-graf.fr infura.io inherit.live initiativesforchina.org inkbunny.net inkui.com inmediahk.net inoreader.com inote.tw insecam.org inside.com.tw insidevoa.com instagram.com instanthq.com institut-tibetain.org interactivebrokers.com internet.org internetdefenseleague.org internetfreedom.org internetpopculture.com inthenameofconfuciusmovie.com investigating.wordpress.com invidio.us inxian.com iownyour.org ipdefenseforum.com ipfire.org ipfs.4everland.io ipfs.io iphone4hongkong.com iphonetaiwan.org iphonix.fr ipicture.ru ipify.org ipjetable.net ipobar.com ipoock.com iportal.me ippotv.com ipredator.se iptv.com.tw iptvbin.com ipvanish.com irangov.ir iredmail.org irna.ir ironpython.net ironsocket.com is-a-hunter.com is.gd isaacmao.com isasecret.com isc.sans.edu isgreat.org ishr.ch islahhaber.net islam.org.hk islamawareness.net islamhouse.com islamicity.com islamicpluralism.org islamtoday.net ismaelan.com ismalltits.com ismprofessional.net isohunt.com israbox.com issuu.com istars.co.nz istockphoto.com isunaffairs.com isuntv.com isupportuyghurs.org itaiwan.gov.tw italiatibet.org itasoftware.com itch.io itemdb.com itemfix.com ithelp.ithome.com.tw itiger.com itsaol.com itshidden.com itsky.it itweet.net iu45.com iuhrdf.org iuksky.com ivacy.com ivonblog.com ivpn.net iwara.tv ixquick.com ixxx.com iyouport.com iyouport.org izaobao.us izles.net izlesem.org j.mp jable.tv jamaat.org jamestown.org jamyangnorbu.com jan.ai japan-whores.com japanhdv.com japantimes.co.jp jav.com jav101.com jav321.com jav68.tv javakiba.org javbus.com javdb.com javfinder.ai javfor.me javhd.com javhip.com javhub.net javhuge.com javlibrary.com javmobile.net javmoo.com javmoo.xyz javseen.com javtag.com javzoo.com jbtalks.cc jbtalks.com jbtalks.my jcpenney.com jdwsy.com jeanyim.com jetos.com jex.com jgoodies.com jiangweiping.com jiaoyou8.com jichangtj.com jiehua.cz jieshibaobao.com jifangge.com jigglegifs.com jihadintel.meforum.org jihadology.net jiji.com jims.net jingpin.org jingsim.org jinpianwang.com jinrizhiyi.news jintian.net jinx.com jitouch.com jjgirls.com jkb.cc jkforum.net jkub.com jma.go.jp jmsc.hku.hk jmscult.com joachims.org jobnewera.wordpress.com joinclubhouse.com joinmastodon.org jornaldacidadeonline.com.br journalchretien.net journalofdemocracy.org joymiihub.com joyourself.com jp.hao123.com jpl.nasa.gov jpopforum.net jsdelivr.net jtvnw.net judicial.gov.tw juhuaren.com jukujo-club.com juliepost.com juliereyc.com junauza.com june4commemoration.org jungleheart.com juoaa.com justdied.com justfreevpn.com justhost.ru justmysocks.net justmysocks1.net justmysockscn.com justpaste.it justtristan.com juziyue.com jwmusic.org jyxf.net ka-wai.com kadokawa.co.jp kagyu.org kagyu.org.za kagyumonlam.org kagyunews.com.hk kagyuoffice.org kagyuoffice.org.tw kaiyuan.de kakao.com kanald.com.tr kankan.today kannewyork.com kanshifang.com kantie.org kanzhongguo.com kanzhongguo.eu kaotic.com karayou.com karmapa-teachings.org karmapa.org kawaiikawaii.jp kawase.com kb.monitorware.com kba-tx.org kcoolonline.com kebrum.com kechara.com keepandshare.com keezmovies.com kendatire.com kenengba.com kepard.com kex.com keycdn.com khabdha.org khatrimaza.org kichiku-doujinko.com kik.com killwall.com kindle4rss.com kindleren.com kineox.free.fr kingdomsalvation.org kinghost.com kingkong.com.tw kingstone.com.tw kink.com kinmen.org.tw kinmen.travel kinokuniya.com kir.jp kissbbao.cn kiwi.kz kk-whys.co.jp kkbox.com kknews.cc kmuh.org.tw knowledgerush.com knowyourmeme.com kobo.com kobobooks.com kodingen.com kompozer.net konachan.com kone.com koolsolutions.com koornk.com koranmandarin.com korenan2.com kpkuang.org kqes.net kraken.com ksdl.org ksnews.com.tw kspcoin.com ktzhk.com kuaichedao.co kucoin.com kui.name kukuku.uk kun.im kurashsultan.com kurtmunger.com kusocity.com kwcg.ca kwongwah.com.my kxsw.life kyofun.com kyohk.net kzaobao.com kzeng.info la-forum.org lab.skk.moe labiennale.org ladbrokes.com lagranepoca.com lala.im lalulalu.com lama.com.tw lamayeshe.com lamnia.co.uk lamrim.com landofhope.tv lanterncn.cn lantosfoundation.org laod.cn laogai.org laogairesearch.org laomiu.com laoyang.info laqingdan.net larsgeorge.com lastcombat.com lastfm.es latelinenews.com lausan.hk lbank.info ld.hao123img.com ldplayer.net ldplayer.tw le-vpn.com leafyvpn.net lecloud.net ledger.com leeao.com.cn lefora.com left21.hk legalporno.com legra.ph legsjapan.com leisurecafe.ca leisurepro.com lematin.ch lemonde.fr lenwhite.com lesoir.be letou.com letscorp.net lflink.com lflinkup.com lflinkup.net lflinkup.org lhakar.org lhasocialwork.org li.taipei liangyou.net liangzhichuanmei.com lianyue.net liaowangxizang.net liberal.org.hk libertysculpturepark.com libertytimes.com.tw library.usc.cuhk.edu.hk libredd.it lifemiles.com lighten.org.tw lighti.me lightnovel.cn lightyearvpn.com lih.kg lihkg.com like.com lilaoshibushinilaoshi.com limiao.net line-apps.com line-scdn.net line.me line.naver.jp linear-abematv.akamaized.net linglingfa.com lingualeo.com lingvodics.com link-o-rama.com linkedin.com linkideo.com linktr.ee linkuswell.com linux.org.hk linuxtoy.org lionsroar.com lipuman.com liquiditytp.com liquidvpn.com listennotes.com listentoyoutube.com listorious.com lists.w3.org litenews.hk liu-xiaobo.org liuhanyu.com liuxiaobo.net liuxiaotong.com livecoin.net livedoor.jp liveleak.com livemint.com livestream.com livevideo.com livingstream.com lizhizhuangbi.com lkcn.net load.to lobsangwangyal.com localbitcoins.com localdomain.ws localpresshk.com lockestek.com login.target.com logos.com.hk londonchinese.ca longhair.hk longmusic.com longtermly.net longtoes.com lookpic.com looktoronto.com lotsawahouse.org lotuslight.org.tw lovetvshow.com lpsg.com lrfz.com lrip.org lsd.org.hk lsforum.net lsm.org lsmchinese.org lsmkorean.org lsmradio.com lsmwebcast.com lsxszzg.com ltn.com.tw luckydesigner.space luckymobile.ca ludepress.com luke54.com luke54.org lupm.org lushstories.com luxebc.com lvhai.org lvv2.com ly.gov.tw lyfhk.net lzjscript.com lzmtnews.org m-team.cc m.hkgalden.com m.me m.moegirl.org ma.hao123.com macgamestore.com macrovpn.com mad-ar.ch madewithcode.com madonna-av.com madou.club madrau.com madthumbs.com mahabodhi.org mahjongsoul.com maiio.net mail-archive.com main-ecnpaper-economist.content.pugpig.com maiplus.com maizhong.org makemymood.com makkahnewspaper.com malaysiakini.com mamingzhe.com manchukuo.net mandiant.com mangabz.com mangafox.com mangafox.me mangmang.run maniash.com manicur4ik.ru mansion.com mansionpoker.com manta.com manyvoices.news maplew.com marc.info marguerite.su martau.com martsangkagyuofficial.org maruta.be marxist.com marxist.net marxists.org mash.to maskedip.com mastodon.cloud mastodon.host mastodon.online mastodon.social mastodon.xyz matainja.com material.io matome-plus.com matome-plus.net matrix.org matters.news matters.town matthewdgreen.wordpress.com mattwilcox.net maven.neoforged.net maxing.jp mayimayi.com mcadforums.com mcaf.ee mcfog.com mcreasite.com mcusercontent.com md-t.org me.me me.ns.ci me.youthwant.com.tw meansys.com media.nu.nl media.org.hk mediachinese.com mediafire.com mediafreakcity.com mediawiki.org medium.com meetav.com meetup.com mefeedia.com mefound.com mega.co.nz mega.io mega.nz megalodon.jp megaproxy.com megurineluka.com meizhong.blog meizhong.report meltoday.com memehk.com memes.tw memri.org memrijttm.org mercari.com mercari.jp mercatox.com mercdn.net mercyprophet.org mergersandinquisitions.com meridian-trust.org meripet.com merit-times.com.tw mesotw.com messenger.com meta.com metafilter.com metamask.io metart.com metarthunter.com meteorshowersonline.com metro.taipei metrohk.com.hk metrolife.ca metroradio.com.hk mewe.com mfxmedia.com mgoon.com mgstage.com mh4u.org mhradio.org microvpn.com middle-way.net mihr.com mihua.org mikanani.me mike.cz.cc mikesoltys.com milph.net milsurps.com mimiai.net mimivip.com mindrolling.org mingdemedia.org minghui-school.org minghui.or.kr minghui.org minghuiyw.wordpress.com mingjinglishi.com mingjingnews.com mingjingtimes.com mingpao.com mingpaocanada.com mingpaomonthly.com mingpaonews.com mingpaony.com mingpaosf.com mingpaotor.com mingpaovan.com mingshengbao.com minhhue.net miningpoolhub.com ministrybooks.org minzhuzhongguo.org miraheze.org miroguide.com mirror.xyz mirrorbooks.com mirrormedia.mg missav.com missav.ws mist.vip mitbbs.com miuipolska.pl mixero.com mixi.jp mixpod.com mixx.com mizzmona.com mjib.gov.tw mjlsh.usc.cuhk.edu.hk mlc.ai mlcool.com mlzs.work mm-cg.com mmaaxx.com mmmca.com mnewstv.com mo.nightlife141.com mobatek.net mobile01.com mobileways.de moby.to mobypicture.com mod.io modernchinastudies.org moeaic.gov.tw moeerolibrary.com moeshare.cc mofa.gov.tw mofos.com mog.com mohu.club mohu.rocks moj.gov.tw mojim.com mol.gov.tw molihua.org momoshop.com.tw mondex.org money-link.com.tw moneydj.com moneyhome.biz monica.im monitorchina.org monocloud.me monster.com moodyz.com moon.fm moonbbs.com moonbingo.com moptt.tw moresci.sale morningsun.org mos.ru motherless.com motor4ik.ru mousebreaker.com movements.org moviefap.com mp3buscador.com mpettis.com mpfinance.com mpinews.com mponline.hk mrbasic.com mrbonus.com mrface.com mrslove.com mrtweet.com msa-it.org msguancha.com msha.gov mstdn.social mswe1.org mthruf.com mubi.com muchosucko.com mullvad.net multiply.com multiproxy.org multiupload.com mummysgold.com musicade.net musixmatch.com muslimvideo.com muzi.com muzi.net muzu.tv mvdis.gov.tw mvg.jp mx.hao123.com mx981.com my-formosa.com my-private-network.co.uk my-proxy.com my.mail.ru my.opera.com my.pcloud.com my03.com myactimes.com myaudiocast.com myav.com.tw mybbs.us mybet.com myca168.com mycanadanow.com mychinanet.com mychinanews.com mychinese.news mycnnews.com mycould.com mydad.info myddns.com myeasytv.com myeclipseide.com myfreecams.com myfreepaysite.com myfreshnet.com myftp.info myiphide.com myjs.tw mykomica.org mylftv.com mymoe.moe mymom.info mynetav.net mynetav.org mynumber.org myparagliding.com mypicture.info mypikpak.com mypop3.net mypop3.org mypopescu.com myreadingmanga.info mysecondarydns.com myshare.url.com.tw mysinablog.com mysite.verizon.net myspace.com myspacecdn.com mytalkbox.com mytizi.com mywife.cc mywww.biz myz.info naacoalition.org naitik.net nakedsecurity.sophos.com nakido.com nakuz.com nalandabodhi.org nalandawest.org namgyal.org namgyalmonastery.org nanopool.org nanyang.com nanyangpost.com nanzao.com naol.ca naol.cc nat.gov.tw nat.moe national-lottery.co.uk nationalawakening.org nationalinterest.org nationalreview.com nationsonline.org nationwide.com naughtyamerica.com naver.com navyfamily.navy.mil navyreserve.navy.mil naweeklytimes.com nbcnews.com nbtvpn.com nbyy.tv nccwatch.org.tw nch.com.tw nchrd.org ncn.org nde.de ndi.org ndr.de ned.org nekoslovakia.net nemesis2.qx.net neo-miracle.com neowin.net netalert.me netbirds.com netcolony.com netflav.com netflix.com netflix.net netlify.app netme.cc netsarang.com netsneak.com network54.com networkedblogs.com networktunnel.net new-3lunch.net new-akiba.com new96.ca newcenturymc.com newcenturynews.com newchen.com newgrounds.com newhighlandvision.com newindianexpress.com newipnow.com newlandmagazine.com.au newmitbbs.com newnews.ca news.cnyes.com news.ebc.net.tw news.msn.com.tw news.mt.co.kr news.nationalgeographic.com news.seehua.com news.sina.com.hk news.sinchew.com.my news.singtao.ca news.tvbs.com.tw news.ycombinator.com news1.kr news100.com.tw news18.com newsancai.com newsblur.com newschinacomment.org newsdetox.ca newsdh.com newsmax.com newstamago.com newstapa.org newstarnet.com newstatesman.com newsweek.com newtaiwan.com.tw newtalk.tw newthuhole.com newyorker.com newyorktimes.com nexon.com next11.co.jp nextdigital.com.hk nextmag.com.tw nextmedia.com nexton-net.jp nexttv.com.tw nf.id.au nfjtyd.com nflxext.com nflximg.com nflximg.net nflxso.net nflxvideo.net nftstorage.link nga.mil ngensis.com nhentai.net nhi.gov.tw nhk-ondemand.jp nic.cz.cc nic.gov nicovideo.jp nightswatch.top nikke-en.com nikke-jp.com nikke-kr.com nikke.hotcool.tw nikkei.com ninecommentaries.com ninjacloak.com ninjaproxy.ninja nintendium.com nirsoft.net nitter.cc nitter.net niu.moe niusnews.com njactb.org nko.navy.mil nlfreevpn.com nmsl.website nnews.eu no-ip.org nobel.se nobelprize.org nodeseek.com nokogiri.org nokola.com noodlevpn.com norbulingka.org nordstrom.com nordstromimage.com nordstromrack.com nordvpn.com nos.nl notepad-plus-plus.org nottinghampost.com now.com now.im nownews.com nowtorrents.com noxinfluencer.com npa.go.jp npa.gov.tw npm.gov.tw npnt.me npsboost.com nradio.me nrk.no ns01.biz ns01.info ns01.us ns02.biz ns02.info ns02.us ns1.name ns2.name ns3.name nsc.gov.tw ntbk.gov.tw ntbna.gov.tw ntbt.gov.tw ntd.tv ntdtv.ca ntdtv.co.kr ntdtv.com ntdtv.com.tw ntdtv.cz ntdtv.org ntdtv.ru ntdtvla.com ntrfun.com nubiles.net nuexpo.com nukistream.com nurgo-software.com nusatrip.com nutaku.net nutsvpn.work nuuvem.com nuvid.com nvdst.com nvquan.org nvtongzhisheng.org nwtca.org nyaa.eu nyaa.si nybooks.com nylon-angel.com nylonstockingsonline.com nypost.com nyt.com nytchina.com nytcn.me nytco.com nyti.ms nytimes.com nytimes.map.fastly.net nytimg.com nytstyle.com nzchinese.com o3o.ca oanda.com oann.com observechina.net obutu.com obyte.org ocaspro.com occupytiananmen.com ocreampies.com ocry.com october-review.org oculus.com oculuscdn.com odysee.com oex.com officeoftibet.com ofile.org ogaoga.org ogate.org ohmyrss.com oikos.com.tw oiktv.com ok.ru okayfreedom.com okex.com okk.tw okpool.me okx.com old.honeynet.org olehdtv.com olelive.com olevod.com olumpo.com olympicwatch.org omct.org omgili.com omni7.jp omnitalk.com omnitalk.org omny.fm on.cc on2.com onapp.com one.one.one.one onedrive.com onedrive.live.com onedumb.com onejav.com onevps.com onion.city onion.ly online.recoveryversion.org onlinecha.com onlineyoutube.com onlygayvideo.com onlytweets.com onmoon.com onmoon.net onmypc.biz onmypc.info onmypc.net onmypc.org onthehunt.com ontrac.com oojj.de open.com.hk open.firstory.me openai.com opendemocracy.net opendn.xyz openid.net openleaks.org opensea.io openstreetmap.org opentech.fund openvpn.net openvpn.org openwebster.com openwrt.org.cn opml.radiotime.com opus-gaming.com ordns.he.net organcare.org.tw organharvestinvestigation.net organiccrap.com orgasm.com orgfree.com oricon.co.jp orient-doll.com orientaldaily.com.my orn.jp osfoora.com otcbtc.com otto.de otzo.com ourdearamy.com ourhobby.com oursogo.com oursteps.com.au oursweb.net ourtv.hk overcast.fm overdaily.org overplay.net oversea.istarshine.com ovpn.com ow.ly owl.li owltail.com oxfordscholarship.com oyax.com oyghan.com ozchinese.com ozvoice.org ozxw.com ozyoyo.com pachosting.com pacificpoker.com packetix.net pacom.mil pacopacomama.com padmanet.com page.link page2rss.com paimon.moe palacemoon.com paljorpublications.com paltalk.com panamapapers.sueddeutsche.de pancakeswap.finance pandafan.pub pandapow.co pandapow.net pandavpn-jp.com pandavpnpro.com pandora.com pandora.tv panoramio.com pao-pao.net paper.li paperb.us paradisehill.cc paradisepoker.com parkansky.com parler.com parse.com parsevideo.com partycasino.com partypoker.com passion.com passiontimes.hk paste.ee pastebin.com pastie.org patreon.com patreonusercontent.com pawoo.net paxful.com payments-jp.amazon.com pbs.org pbwiki.com pbworks.com pbxes.com pbxes.org pcanywhere.net pcc.gov.tw pcdvd.com.tw pcgamestorrents.com pchome.com.tw pcij.org pcstore.com.tw pct.org.tw pdetails.com pdproxy.com pds.nasa.gov peace.ca peacefire.org peeasian.com peing.net pekingduck.org pemulihan.or.id pen.io penchinese.com pendrivelinux.com penthouse.com pentoy.hk peoplebookcafe.com peoplenews.tw peopo.org percy.in perfect-privacy.com perfectgirls.net periscope.tv perplexity.ai persecutionblog.com persiankitty.com pewresearch.org phapluan.org phayul.com philborges.com phmsociety.org phncdn.com phosphation13.rssing.com photodharma.net photofocus.com photonmedia.net piaotia.com picacomic.com picacomiccn.com picasaweb.com picidae.net picturedip.com picturesocial.com picuki.com pigav.com pimg.tw pin-cong.com pin6.com pincong.rocks ping.fm pinimg.com pinkrod.com pinoy-n.com pioneer-worker.forums-free.com pipii.tv piraattilahti.org piring.com pixeldrain.com pixelqi.com pixiv.net pixiv.org pixivsketch.net pixnet.net pk.com pki.goog pkqjiasu.com pkuanvil.com placemix.com play-asia.com playboy.com playboyplus.com player.fm playno1.com playpcesor.com plexvpn.pro plm.org.hk plunder.com plurk.com plus.codes plus28.com plusbb.com pmatehunter.com pmates.com po2b.com pobieramy.top podbean.com podcast.co podictionary.com poe.com points-media.com pokerstars.com pokerstars.net politicalchina.org politiscales.net poloniex.com polymarket.com polymer-project.org polymerhk.com poolbinance.com poolin.com popai.pro popo.tw popvote.hk popxi.click popyard.com popyard.org porn.com porn2.com porn5.com pornbase.org pornerbros.com pornhd.com pornhost.com pornhub.com pornhubdeutsch.net pornmate.com pornoxo.com pornrapidshare.com pornsharing.com pornsocket.com pornstarbyface.com pornstarclub.com porntube.com porntubenews.com porntvblog.com pornvisit.com port25.biz portablevpn.nl poskotanews.com post01.com post76.com post852.com postadult.com posts.careerengine.us potato.im potvpn.com pourquoi.tw power.com powercx.com powerphoto.org ppy.sh prayforchina.net prcleader.org premproxy.com presentation.new presentationzen.com president.ir presidentlee.tw prestige-av.com primevideo.com prism-break.org prisoneralert.com pritunl.com privacybox.de privacyguides.org private.com privateinternetaccess.com privatepaste.com privatetunnel.com privatevpn.com privoxy.org procopytips.com project-syndicate.org prosiben.de proton.me protonvpn.com provideocoalition.com provpnaccounts.com proxfree.com proxifier.com proxomitron.info proxpn.com proxyanonimo.es proxydns.com proxynetwork.org.uk proxyroad.com proxytunnel.net proxz.com proyectoclubes.com pscp.tv pshvpn.com psiphon.ca psiphon3.com psiphontoday.com pstatic.net pt.im ptt.cc pttgame.com pttvan.org ptwxz.com pubu.com.tw puffinbrowser.com puffstore.com pullfolio.com punyu.com pure18.com pureconcepts.net puredns.org pureinsight.org purepdf.com purevpn.com purplelotus.org purpose.nike.com pursuestar.com pussyspace.com putihome.org putlocker.com putty.org pwned.com pximg.net python.com python.com.tw pythonhackers.com pytorch.org qbittorrent.org qgirl.com.tw qhigh.com qi-gong.me qianbai.tw qiandao.today qianglie.com qiangwaikan.com qiangyou.org qianmo.tw qidian.ca qiwen.lu qixianglu.cn qkshare.com qmp4.com qoos.com qpoe.com qq.co.za qstatus.com qtrac.eu questvisual.com quitccp.org quiz.directory quora.com quoracdn.net quran.com quranexplorer.com qusi8.net qxbbs.org qz.com r-pool.net r0.ru r10s.jp r18.com radicalparty.org radiko.jp radio-canada.ca radio-en-ligne.fr radio.garden radioaustralia.net.au radiohilight.net radioline.co radiovaticana.org radiovncr.com radmin-vpn.com rael.org raggedbanner.com raidcall.com.tw rainbowplan.org raindrop.io raizoji.or.jp rakuten.co.jp ramcity.com.au rangzen.net rangzen.org rapbull.net rapidmoviez.com rapidvpn.com rarbgprx.org rateyourmusic.com rationalwiki.org raw.githack.com rawgit.com rawgithub.com rcam.target.com rcinet.ca rconversation.blogs.com rd.com reabble.com read01.com read100.com readingtimes.com.tw readmoo.com readydown.com realcourage.org realforum.zkiz.com realitykings.com realraptalk.com realsexpass.com reason.com rebatesrule.net recordhistory.org recovery.org.tw recoveryversion.com.tw red-lang.org redbubble.com redchinacn.net redchinacn.org redd.it reddit.com reddithelp.com redditlist.com redditmedia.com redditspace.com redditstatic.com redhotlabs.com redtube.com referer.us reflectivecode.com relaxbbs.com relay.com.tw relay.firefox.com releaseinternational.org religionnews.com renminbao.com renyurenquan.org resilio.com resistchina.org retweetist.com retweetrank.com reuters.com reutersmedia.net revleft.com revver.com rfa.org rfachina.com rfalive1.akacast.akamaistream.net rfamobile.org rfaweb.org rferl.org rfi.fr rfi.my rigpa.org riku.me rileyguide.com riseup.net ritouki.jp ritter.vg rixcloud.com rixcloud.us rlwlw.com rmbl.ws rmjdw.com roadshow.hk roboforex.com robustnessiskey.com rocket-inc.net rocket.chat rocksdb.org rojo.com rolfoundation.org rolia.net rolsociety.org ronjoneswriter.com roodo.com rosechina.net rou.video rsdlmonitor.com rsf-chinese.org rsf.org rsshub.app rssmeme.com rtalabel.org rthk.hk rthk.org.hk rthklive2-lh.akamaihd.net rti.org.tw rti.tw rtm.tnt-ea.com ruanyifeng.com rukor.org rule34.xxx rule34video.com rumble.com runbtx.com rushbee.com rusvpn.com ruten.com.tw rutracker.net rutracker.org rutube.ru rxhj.net s-cute.com s-dragon.org s.yimg.com s1s1s1.com s3-ap-northeast-1.amazonaws.com s3-ap-northeast-2.amazonaws.com s3-ap-southeast-1.amazonaws.com s3-ap-southeast-2.amazonaws.com s3-eu-central-1.amazonaws.com s3.amazonaws.com s3.ap-northeast-2.amazonaws.com s3.eu-central-1.amazonaws.com s3.us-east-1.amazonaws.com sa.hao123.com sacks.com sacom.hk sadistic-v.com sadpanda.us safechat.com safeguarddefenders.com safervpn.com sagernet.org saintyculture.com sakuralive.com sakya.org salvation.org.hk samair.ru sambhota.org sankakucomplex.com sankei.com sanmin.com.tw sapikachu.net savemedia.com savethedate.foo savethesounds.info savetibet.de savetibet.fr savetibet.nl savetibet.org savetibet.ru savetibetstore.org saveuighur.org savevid.com sbme.me sbs.com.au scache.vzw.com scache1.vzw.com scache2.vzw.com scasino.com schema.org sciencenets.com scieron.com scmp.com scmpchinese.com scramble.io scratch.mit.edu scribd.com scriptspot.com search.aol.com search.com search.xxx search.yahoo.co.jp searchtruth.com searx.me seattlefdc.com secretchina.com secretgarden.no secretsline.biz secure.hustler.com secure.logmein.com secure.shadowsocks.nu secureservercdn.net securetunnel.com securityinabox.org securitykiss.com seed4.me seesmic.com seevpn.com seezone.net sehuatang.net sehuatang.org sejie.com sellclassics.com sendsmtp.com sendspace.com sensortower.com servehttp.com serveuser.com serveusers.com sesawe.net sethwklein.net setn.com settv.com.tw sevenload.com sex.com sex3.com sex8.cc sexandsubmission.com sexbot.com sexhu.com sexidude.com sexinsex.net sextvx.com sf.net sfileydy.com sfshibao.com sftindia.org sftuk.org shadeyouvpn.com shadow.ma shadowsky.xyz shadowsocks.asia shadowsocks.be shadowsocks.com shadowsocks.com.hk shadowsocks.org shahit.biz shambalapost.com shapeservices.com share-videos.se share.america.gov share.ovi.com share.youthwant.com.tw sharebee.com sharecool.org sharpdaily.hk sharpdaily.tw shat-tibet.com shattered.io sheet.new sheets.new sheikyermami.com shellfire.de shenyun.com shenyunperformingarts.org shenyunshop.com shenzhoufilm.com shenzhouzhengdao.org shiatv.net shicheng.org shiksha.com shipcamouflage.com shireyishunjian.com shitaotv.org shixiao.org shizhao.org shkspr.mobi shodanhq.com shooshtime.com shop2000.com.tw shopee.tw shopping.com shopping.yahoo.co.jp showhaotu.com showtime.jp showwe.tw shutterstock.com shwchurch.org shwchurch3.com siddharthasintent.org sidelinesnews.com sidelinessportseatery.com sierrafriendsoftibet.org signal.org sijihuisuo.club sijihuisuo.com silkbook.com silvergatebank.com simbolostwitter.com simplecd.me simplecd.org simpleproductivityblog.com simpleswap.io simplex.chat sina.com.hk sinchew.com.my singaporepools.com.sg singfortibet.com singlelogin.se singpao.com.hk singtao.com singtaousa.com sino-monthly.com sinoants.com sinoca.com sinocast.com sinocism.com sinoinsider.com sinomontreal.ca sinoquebec.com sipml5.org sis.xxx sis001.com sis001.us site.new site2unblock.com sitebro.tw sitekreator.com sitemaps.org sites.new six-degrees.io sketchappsources.com skimtube.com skybet.com skyking.com.tw skyvegas.com skyxvpn.com slacker.com slashine.onl slaytizle.com sleazydream.com sleazyfork.org slheng.com slickvpn.com slides.com slides.new slideshare.net slinkset.com slutload.com slutmoonbeam.com slyip.com slyip.net sm-miracle.com smartdnsproxy.com smarthide.com smchbooks.com smh.com.au smhric.org smith.edu smn.news smyxy.org snapseed.com sndcdn.com sneakme.net snowlionpub.com sobees.com soc.mil social.datalabour.com social.edu.ci socialblade.com socks-proxy.net sockscap64.com sockslist.net socrec.org sod.co.jp softether-download.com softether.co.jp softether.org softfamous.com softwarebychuck.com softwaredownload.gitbooks.io sogclub.com sogrady.me soh.tw sohcradio.com sohfrance.org sokamonline.com sokmil.com solana.com solarsystem.nasa.gov solidaritetibet.org solidfiles.com solv.finance somee.com songjianjun.com sonidodelaesperanza.org sopcast.com sopcast.org sorting-algorithms.com sos.org sosad.fun sosreader.com soubory.com soul-plus.net soulcaliburhentai.net soundcloud.com soundofhope.kr soundofhope.org soundon.fm soup.io sourceforge.net sourcewadio.com south-plus.net south-plus.org southmongolia.org southnews.com.tw southpark.cc.com sowers.org.hk soylentnews.org spaces.hightail.com spankbang.com spankingtube.com spankwire.com sparkpool.com spatial.io spb.com speakerdeck.com specxinzl.jigsy.com speedcat.me speedify.com spencertipping.com spendee.com spicevpn.com spideroak.com spiderpool.com spike.com sports.williamhill.com spotflux.com spotify.com spreadsheet.new spreadshirt.es spreaker.com spring4u.info springboardplatform.com springwood.me sprite.org sproutcore.com squirly.info squirrelvpn.com srcf.ucam.org ss-link.com ss.pythonic.life ss7.vzw.com ssglobal.co ssglobal.me ssl.webpack.de ssl443.org sspanel.net ssr.tools ssrshare.com ssrshare.us ssrtool.com sstm.moe sstmlt.moe sstmlt.net stackoverflow.com standard.co.uk standupfortibet.org standwithhk.org stanford.edu starfishfx.com starp2p.com startpage.com startuplivingchina.com stat.gov.tw static-economist.com static.shemalez.com static01.nyt.com staticflickr.com stboy.net stc.com.sa steamcommunity.com steamstatic.com steel-storm.com steemit.com steganos.com steganos.net stepchina.com stephaniered.com sthoo.com stickam.com stickeraction.com stileproject.com stitcher.com sto.cc stoporganharvesting.org stoptibetcrisis.net storage.yandex.net storagenewsletter.com store.steampowered.com storj.io storm.mg stormmediagroup.com storry.tv stoweboyd.com straitstimes.com stranabg.com straplessdildo.com streamable.com streamate.com streamingthe.net streema.com strikingly.com strongvpn.com strongwindpress.com studentsforafreetibet.org stumbleupon.com stupidvideos.com subhd.tv substack.com successfn.com suche.gmx.net sugarsync.com sugobbs.com sugumiru18.com suissl.com sujiatun.wordpress.com summify.com sumrando.com sun1911.com sundayguardianlive.com sunmedia.ca suno.ai suno.com sunporno.com sunskyforum.com sunta.com.tw sunvpn.net sunwinism.joinbbs.net supchina.com superfreevpn.com superpages.com supervpn.net superzooi.com suppig.net suprememastertv.com surfeasy.com surfeasy.com.au surfshark.com suroot.com surrenderat20.net svsfx.com swagbucks.com swapspace.co swissinfo.ch swissvpn.net switch1.jp switchvpn.net sydneytoday.com sylfoundation.org synapse.org syncback.com synergyse.com syosetu.com sysresccd.org sytes.net szbbs.net szetowah.org.hk t-g.com t.co t.me t35.com t66y.com taa-usa.org taaze.tw tabtter.jp tacc.cwb.gov.tw taconet.com.tw taedp.org.tw tafm.org tagwalk.com tahr.org.tw taipei.gov.tw taipeisociety.org taipeitimes.com taisounds.com taiwanbible.com taiwandaily.net taiwandc.org taiwanhot.net taiwanjobs.gov.tw taiwanjustice.com taiwanjustice.net taiwankiss.com taiwannation.50webs.com taiwannation.com taiwannation.com.tw taiwanncf.org.tw taiwannews.com.tw taiwantp.net taiwantt.org.tw taiwanus.net taiwanyes.ning.com talk853.com talkatone.com talkboxapp.com talkcc.com talkonly.net tanc.org tangren.us tanks.gg taoism.net tapanwap.com tapatalk.com tardigrade.io tarr.uspto.gov taup.net taweet.com tbcollege.org tbi.org.hk tbjyt.org tbrc.org tbs-rainbow.org tbsec.org tbskkinabalu.page.tl tbsn.org tbsseattle.org tbssqh.org tbswd.org tbtemple.org.uk tbthouston.org tccwonline.org tcewf.org tchrd.org tcnynj.org tcpspeed.co tcsofbc.org tdm.com.mo teachparentstech.org teamamericany.com technews.tw techspot.com techviz.net teck.in teco-hk.org teco-mo.org teddysun.com teeniefuck.net teensinasia.com tehrantimes.com telecomspace.com telega.one telegra.ph telegram.dog telegram.me telegram.org telegram.space telegramdownload.com telegraph.co.uk telesco.pe tellapart.com tellme.pw tenacy.com tenor.com tensorflow.org tenzinpalmo.com terabox.com tew.org textnow.com textnow.me tfc-taiwan.org.tw tfhub.dev tfiflve.com tg-me.com tg.dev th.hao123.com thaicn.com thb.gov.tw theatlantic.com theatrum-belli.com thebcomplex.com theblaze.com theblemish.com thebobs.com thebodyshop-usa.com thecenter.mit.edu thechasernews.co.uk thechinabeat.org thechinacollection.org theconversation.com thedalailamamovie.com thediplomat.com thedw.us theepochtimes.com thefacebook.com thegay.com thegioitinhoc.vn thegly.com theguardian.com thehansindia.com thehindu.com thehun.net theinitium.com thenewslens.com thepiratebay.org theporndude.com theportalwiki.com theprint.in therock.net.nz thesaturdaypaper.com.au thestandnews.com thetatoken.org thetibetcenter.org thetibetconnection.org thetibetmuseum.org thetibetpost.com thetrotskymovie.com thetvdb.com thewgo.org thewirechina.com theync.com thinkgeek.com thinkingtaiwan.com thinkwithgoogle.com thirdmill.org thisav.com thlib.org thomasbernhard.org thongdreams.com threadreaderapp.com threads.com threads.net throughnightsfire.com thuhole.com thumbzilla.com thywords.com tiananmenduizhi.com tiananmenmother.org tiananmenuniv.com tiananmenuniv.net tiandixing.org tianhuayuan.com tianlawoffice.com tianti.io tiantibooks.org tianyantong.org.cn tianzhu.org tibet-envoy.eu tibet-foundation.org tibet-house-trust.co.uk tibet-initiative.de tibet-munich.de tibet.at tibet.ca tibet.com tibet.fr tibet.net tibet.nu tibet.org tibet.org.tw tibet.to tibet3rdpole.org tibetaction.net tibetaid.org tibetalk.com tibetan-alliance.org tibetan.fr tibetanaidproject.org tibetanarts.org tibetanbuddhistinstitute.org tibetancommunity.org tibetancommunityuk.net tibetanculture.org tibetanentrepreneurs.org tibetanfeministcollective.org tibetanhealth.org tibetanjournal.com tibetanlanguage.org tibetanliberation.org tibetanpaintings.com tibetanphotoproject.com tibetanpoliticalreview.org tibetanreview.net tibetansports.org tibetanwomen.org tibetanyouth.org tibetanyouthcongress.org tibetcharity.dk tibetcharity.in tibetchild.org tibetcity.com tibetcollection.com tibetcorps.org tibetexpress.net tibetfocus.com tibetfund.org tibetgermany.com tibetgermany.de tibethaus.com tibetheritagefund.org tibethouse.jp tibethouse.org tibethouse.us tibetinfonet.net tibetjustice.org tibetkomite.dk tibetlibre.free.fr tibetmuseum.org tibetnetwork.org tibetoffice.ch tibetoffice.com.au tibetoffice.eu tibetoffice.org tibetonline.com tibetonline.tv tibetoralhistory.org tibetpolicy.eu tibetrelieffund.co.uk tibetsociety.com tibetsun.com tibetsupportgroup.org tibetswiss.ch tibettelegraph.com tibettimes.net tibettruth.com tibetwrites.org ticket.com.tw tigervpn.com tiktok.com tiktokcdn-eu.com tiktokcdn-us.com tiktokcdn.com tiktokv.com tiktokv.us tiltbrush.com timdir.com time.com timesnownews.com timesofindia.indiatimes.com timsah.com timtales.com tinc-vpn.org tineye.com tingtalk.me tiny.cc tinychat.com tinypaste.com tinyurl.com tipas.net tipo.gov.tw tistory.com tkcs-collins.com tl.gd tma.co.jp tmagazine.com tmi.me tmpp.org tn1.shemalez.com tn2.shemalez.com tn3.shemalez.com tnaflix.com tnp.org to-porno.com togetter.com toh.info token.im tokenlon.im tokyo-247.com tokyo-hot.com tokyo-porn-tube.com tokyocn.com tomp3.cc tongil.or.kr tonyyan.net toonel.net top.tv top10vpn.com top81.ws topbtc.com topic.youthwant.com.tw topnews.in toppornsites.com topshareware.com topsy.com toptip.ca toptoon.net tor.updatestar.com tora.to torcn.com torguard.net torlock.com torproject.org torrentgalaxy.to torrentkitty.tv torrentprivacy.com torrentproject.se torrenty.org tortoisesvn.net torvpn.com tosh.comedycentral.com totalvpn.com tou.tv toutiaoabc.com towngain.com toypark.in toythieves.com toytractorshow.com tparents.org tpi.org.tw tracfone.com tradingview.com translate.goog transparency.org treemall.com.tw trendsmap.com trickip.net trimondi.de tronscan.org trouw.nl trt.net.tr trtc.com.tw truebuddha-md.org trulyergonomic.com truthsocial.com truveo.com tryheart.jp tsctv.net tsdr.uspto.gov tsemtulku.com tsquare.tv tsu.org.tw tsunagarumon.com tt1069.com tttan.com ttv.com.tw ttvnw.net tu8964.com tubaholic.com tube.com tube8.com tube911.com tubecup.com tubegals.com tubeislam.com tubepornclassic.com tubestack.com tubewolf.com tuibeitu.net tuidang.org tuidang.se tuitwit.com tukaani.org tumblr.com tumutanzi.com tumview.com tunein.com tunein.streamguys1.com tunnelbear.com tunnelblick.net tunnelr.com tunsafe.com turansam.org turbobit.net turbohide.com turkistantimes.com turntable.fm tushycash.com tuvpn.com tuzaijidi.com tv.com tv.jtbc.joins.com tvants.com tvboxnow.com tvider.com tvmost.com.hk tvplayvideos.com tvunetworks.com tw-blog.com tw-npo.org tw.gigacircle.com tw.hao123.com tw.jiepang.com tw.streetvoice.com tw.tomonews.net tw01.org twaitter.com twapperkeeper.com twaud.io twavi.com twbbs.org twblogger.com tweepguide.com tweepmag.com tweepml.org tweetbackup.com tweetboard.com tweetcs.com tweetdeck.com tweetedtimes.com tweetphoto.com tweetree.com tweettunnel.com tweetwally.com tweetymail.com tweez.net twelve.today twerkingbutt.com twftp.org twgreatdaily.com twibase.com twibble.de twibbon.com twibs.com twicountry.org twicsy.com twiends.com twifan.com twiffo.com twiggit.org twilightsex.com twilog.org twimbow.com twimg.com twimg.edgesuite.net twip.me twipple.jp twishort.com twister.net.co twisternow.com twistory.net twitch.tv twitchcdn.net twitgoo.com twitiq.com twitlonger.com twitmania.com twitoaster.com twitonmsn.com twitpic.com twitstat.com twittbot.net twitter.com twitter.jp twitter4j.org twittercounter.com twitterfeed.com twittergadget.com twitterkr.com twittermail.com twitterrific.com twittertim.es twitthat.com twitturk.com twitturly.com twitzap.com twiyia.com twkan.com twnorth.org.tw twreporter.org twskype.com twt.tl twtkr.com twtr2src.ogaoga.org twtrland.com twttr.com twurl.nl tx.me txxx.com tycool.com typepad.com typeset.io typora.io u15.info u9un.com ua5v.com ub0.cc ubddns.org uberproxy.net uc-japan.org uchicago.edu udn.com udn.com.tw udnbkk.com udomain.hk uforadio.com.tw ufreevpn.com ugo.com uhdwallpapers.org uhrp.org uighur.narod.ru uighur.nl uighurbiz.net ukcdp.co.uk uku.im ulike.net ulop.net ultrasurf.us ultravpn.com ultravpn.fr ultraxs.com umich.edu unblock-us.com unblock.cn.com unblockdmm.com unblocker.yt unblocksit.es uncyclomedia.org uncyclopedia.hk uncyclopedia.tw underwoodammo.com unholyknight.com uni.cc unification.net unification.org.tw unirule.cloud unix100.com unknownspace.org unmineable.com unodedos.com unpo.org unseen.is unstable.icu unwire.hk uocn.org upbit.com updates.tdesktop.com upghsbc.com upholdjustice.org uploaded.net uploaded.to uploadstation.com upmedia.mg upornia.com uproxy.org uptodown.com upwill.org ur7s.com uraban.me urbandictionary.com urbansurvival.com urchin.com urlborg.com urlparser.com us.to usacn.com usaip.eu uscardforum.com uscg.mil uscnpm.org use.typekit.net usercontent.goog users.skynet.be usfk.mil usma.edu usmgtcg.ning.com usno.navy.mil usocctn.com ustibetcommittee.org ustream.tv usus.cc utopianpal.com uujiasu.com uukanshu.com uupool.cn uvwxyz.xyz uwants.com uwants.net uyghur-j.org uyghur.co.uk uyghuraa.org uyghuramerican.org uyghurbiz.org uyghurcongress.org uyghurpen.org uyghurstudies.org uyghurtribunal.com uygur.fc2web.com uygur.org uymaarip.com v2.help v2ex.com v2fly.org v2ray.com v2raycn.com valeursactuelles.com van001.com van698.com vanemu.cn vanilla-jp.com vanpeople.com vansky.com vaticannews.va vatn.org vcf-online.org vcfbuilder.org vegas.williamhill.com vegasred.com velkaepocha.sk venbbs.com venchina.com venetianmacao.com ventureswell.com veoh.com vercel.app vermonttibet.org vern.cc verybs.com vevo.com vewas.net vft.com.tw viber.com vica.info vid.me vidble.com video.aol.ca video.aol.co.uk video.aol.com video.foxbusiness.com videobam.com videodetective.com videomega.tv videomo.com videopediaworld.com videopress.com vidinfo.org vietdaikynguyen.com vijayatemple.org vilanet.me vilavpn.com vimeo.com vimperator.org vincnd.com vine.co vinniev.com vip-enterprise.com virtualrealporn.com visibletweets.com viu.com viu.tv vivahentai4u.net vivaldi.com vivatube.com vivthomas.com vizvaz.com vjav.com vjmedia.com.hk vllcs.org vmixcore.com vmpsoft.com vn.hao123.com vnet.link voa-11.akacast.akamaistream.net voacambodia.com voacantonese.com voachinese.com voachineseblog.com voagd.com voaindonesia.com voanews.com voatibetan.com voatibetanenglish.com vocaroo.com vocativ.com vocn.tv vocus.cc vod-abematv.akamaized.net vod.wwe.com voicettank.org vot.org vovo2000.com voxer.com voy.com vpl.bibliocommons.com vpn.ac vpn.cmu.edu vpn.net vpn.sv.cmu.edu vpn4all.com vpnaccount.org vpnaccounts.com vpnbook.com vpncomparison.org vpncoupons.com vpncup.com vpndada.com vpnfan.com vpnfire.com vpnforgame.net vpngate.jp vpngate.net vpngratis.net vpnhq.com vpnhub.com vpninja.net vpnintouch.com vpnjack.com vpnmaster.com vpnmentor.com vpnpick.com vpnpop.com vpnpronet.com vpnproxymaster.com vpnreactor.com vpnreviewz.com vpnsecure.me vpnshazam.com vpnshieldapp.com vpnsp.com vpntraffic.com vpntunnel.com vpnuk.info vpnunlimitedapp.com vpnvip.com vpnworldwide.com vporn.com vpser.net vpsxb.net vraiesagesse.net vrchat.com vrmtr.com vrporn.com vrsmash.com vtunnel.com vuku.cc vultryhw.com w-pool.com w.idaiwan.com w3s.link waffle1999.com wahas.com waikeung.org wainao.me walletconnect.com wallmama.com wallpapercasa.com wallproxy.com wallsttv.com waltermartin.com waltermartin.org wanderinghorse.net wangafu.net wangjinbo.org wanglixiong.com wango.org wangruoshui.net want-daily.com wanz-factory.com wapedia.mobi warroom.org waselpro.com washingtonpost.com watch8x.com watchinese.com watchmygf.net watchout.tw wattpad.com wav.tv waveprotocol.org waybig.com waymo.com wd.bible wda.gov.tw wdf5.com wealth.com.tw wearehairy.com wearn.com web.dev web2project.net webbang.net webevader.org webfreer.com webjb.org weblagu.com webmproject.org webpkgcache.com webrtc.org webrush.net webs-tv.net website.informer.com website.new websitepulse.com webwarper.net webworkerdaily.com wechatlawsuit.com weebly.com wefightcensorship.org wefong.com wego.here.com weiboleak.com weihuo.org weijingsheng.org weiming.info weiquanwang.org weisuo.ws welovecock.com welt.de wemigrate.org wengewang.com wengewang.org wenxuecity.com wenyunchao.com wenzhao.ca westca.com westernshugdensociety.org westernwolves.com westkit.net westpoint.edu wetplace.com wetpussygames.com wezone.net wforum.com whatblocked.com whatbrowser.org whats.new whatsapp.com whatsapp.net wheelockslatin.com whereiswerner.com wheretowatch.com whippedass.com whispersystems.org whoer.net whotalking.com whylover.com whyx.org widevine.com wikaba.com wiki.gamerp.jp wiki.jqueryui.com wiki.keso.cn wiki.metacubex.one wiki.oauth.net wiki.phonegap.com wikibooks.org wikidata.org wikileaks-forum.com wikileaks.ch wikileaks.com wikileaks.de wikileaks.eu wikileaks.lu wikileaks.org wikileaks.pl wikiless.funami.tech wikilivres.info wikimapia.org wikimedia.org wikinews.org wikipedia.org wikisource.org wikiversity.org wikivoyage.org wikiwand.com wiktionary.org williamhill.com willw.net wilsoncenter.org windscribe.com wingamestore.com wingy.site winning11.com wionews.com wire.com wiredbytes.com wiredpen.com wireguard.com wisdompubs.org wisevid.com withgoogle.com withyoutube.com witnessleeteaching.com witopia.net wizcrafts.net wjbk.org wlcnew.jigsy.com wmflabs.org wmfusercontent.org wn.com wnacg.com wnacg.org wo.tc wo3ttt.wordpress.com woeser.com wokar.org wolfax.com wombo.ai woolyss.com woopie.jp woopie.tv wordpress.com work2icu.org workatruna.com workerempowerment.org workers.dev worldcat.org worldjournal.com worldvpn.net wow.com wowgirls.com wowhead.com wowporn.com wowrk.com woyaolian.org wozy.in wp.com wpoforum.com wrchina.org wretch.cc writer.zoho.com writesonic.com wsj.com wsj.net wtbn.org wtfpeople.com wuerkaixi.com wufafangwen.com wufi.org.tw wujie.net wujieliulan.com wunderground.com wuw.red wwitv.com www.ajsands.com www.antd.org www.aolnews.com www.businessinsider.com.au www.cmoinc.org www.dmm.com www.dwheeler.com www.eastturkistan.net www.gmiddle.com www.gmiddle.net www.hustlercash.com www.idlcoyote.com www.imdb.com www.kindleren.com www.klip.me www.lib.virginia.edu www.lorenzetti.com.br www.m-sport.co.uk www.monlamit.org www.moztw.org www.msn.com www.nbc.com www.owind.com www.oxid.it www.powerpointninja.com www.s4miniarchive.com www.sciencemag.org www.shadowsocks.com www.skype.com www.tablesgenerator.com www.taiwanonline.cc www.thechinastory.org www.wan-press.org www.websnapr.com www.xicons.org www.zensur.freerk.com www1.american.edu www1.biz www2.ohchr.org www2.rocketbbs.com wwwhost.biz wxw.cat wxw.moe wzyboy.im x-art.com x-berry.com x-wall.org x.ai x.co x.com x.company x24hr.com x3guide.com xanga.com xbabe.com xbookcn.com xbtce.com xcafe.in xcity.jp xcritic.com xerotica.com xfinity.com xfxssr.me xgmyd.com xhamster.com xianba.net xianjian.tw xiaobaiwu.com xiaochuncnjp.com xiaohexie.com xiaolan.me xiaoma.org xiaomi.eu xiaxiaoqiang.net xiezhua.com xihua.es xijie.wordpress.com xing.com xinjiangpolicefiles.org xinmiao.com.hk xinqimeng.over-blog.com xinsheng.net xinshijue.com xiongpian.com xiuren.org xizang-zhiye.org xjp.cc xjtravelguide.com xm.com xml-training-guide.com xmovies.com xn--11xs86f.icu xn--4gq171p.com xn--9pr62r24a.com xn--czq75pvv1aj5c.org xn--i2ru8q2qg.com xn--ngstr-lra8j.com xn--noss43i.com xn--oiq.cc xn--p8j9a0d9c9a.xn--q9jyb4c xnpool.com xnxx.com xpdo.net xpud.org xrentdvd.com xsden.info xskywalker.com xt.com xt.pub xtube.com xuchao.net xuchao.org xuehua.us xvbelink.com xvideo.cc xvideos-cdn.com xvideos.com xvideos.es xvinlink.com xxbbx.com xxlmovies.com xxuz.com xxx.com xxx.xxx xxxfuckmom.com xxxx.com.au xxxy.info xxxymovies.com xys.dxiong.com xys.org xysblogs.org y2mate.com yadi.sk yahoo.com yahoo.com.hk yahoo.com.tw yakbutterblues.com yam.com yam.org.tw yande.re yanghengjun.com yangzhi.org yasni.co.uk yasukuni.or.jp yayabay.com ydy.com yeahteentube.com yecl.net yeelou.com yeeyi.com yegle.net yes-news.com yes.xxx yes123.com.tw yesasia.com yesasia.com.hk yespornplease.com yeyeclub.com ygto.com yhcw.net yibada.com yibaochina.com yidio.com yigeni.com yilubbs.com yinlei.org yipub.com yizhihongxing.com yobit.net yobt.com yobt.tv yogichen.org yolasite.com yomiuri.co.jp yong.hu yorkbbs.ca you-get.org you.com youdontcare.com youjizz.com youmaker.com youngpornvideos.com youngspiration.hk youpai.org youporn.com youporngay.com your-freedom.net yourepeat.com yourlisten.com yourlust.com yourtrap.com yousendit.com youthnetradio.org youtu.be youtube-nocookie.com youtube.com youtubeeducation.com youtubegaming.com youtubekids.com youversion.com youwin.com youxu.info yt.be ytht.net ytimg.com ytn.co.kr yuanming.net yuanzhengtang.org yulghun.com yunchao.net yunomi.tokyo yuvutu.com yvesgeleyn.com ywpw.com yx51.net yyii.org yyjlymb.xyz yysub.net yzzk.com z-lib.fm z-lib.fo z-lib.gd z-lib.gl z-lib.io z-lib.org z-library.sk zacebook.com zalmos.com zamimg.com zaobao.com.sg zapto.org zattoo.com zb.com zdnet.com.tw zello.com zengjinyan.org zenmate.com zenmate.com.ru zerohedge.com zeronet.io zfreet.com zh-hans.cfsh99.com zh.ecdm.wikia.com zh.pokerstrategy.com zh.pttpedia.wikia.com zh.uncyclopedia.wikia.com zh.wikiquote.org zhangboli.net zhangtianliang.com zhanlve.org zhao.1984.city zhao.jinhai.de zhenghui.org zhengjian.org zhengwunet.org zhenxiang.biz zhizhu.top zhongguo.ca zhongguorenquan.org zhongguotese.net zhongzidi.com zhoushuguang.com zhuanxing.cn zhuatieba.com zhuichaguoji.org zi.media ziddu.com zillionk.com zim.vn zinio.com ziporn.com zippyshare.com zmedia.com.tw zmw.cn zodgame.us zodgame.xyz zomobo.net zonaeuropa.com zonghexinwen.com zoogvpn.com zoominfo.com zooqle.com zootool.com zoozle.net zophar.net zorrovpn.com zozotown.com zpn.im zspeeder.me zsrhao.com zuo.la zuobiao.me zuola.com zvereff.com zynamics.com zyns.com zyxel.com zzcartoon.com zzcloud.me zzux.com
297854895/vue-tsx-admin
3,284
README.md
在线示例 ---- [在线预览](http://115.126.79.121:9999/) 页面示例完成情况 ---- - [x] 登录页 - [x] 首页 - [ ] 表单页 - [ ] 基础表单 - [ ] 分步表单 - [ ] 高级表单 - [ ] 列表页 - [x] 查询表格 - [ ] 标准列表 - [ ] 卡片列表 - [ ] 搜索列表 - [ ] 详情页 - [ ] 标准详情 - [ ] 高级详情 - [ ] 个人页 - [ ] 个人中心 - [ ] 个人设置 - [ ] 开发引导示例 - [ ] 主题色 - [ ] Tab组件 - [ ] 整体风格 - [ ] 其他设置 - [ ] 路由动画 - [x] 数据持久化 - [x] 系统数据持久化 - [x] 路由数据持久化 - [ ] 权限管理 - [ ] 登录验证 - [ ] 创建路由 - [ ] 创建菜单 - [x] 404页 主体完成情况 ---- - [x] 登录 - [ ] 注册 - [x] 退出登录 - [x] 登录验证 - [x] 权限创建菜单 - [x] 权限验证路由合法性 - [x] 系统样式自定义 - [x] 响应式布局 - [x] 数据持久化 - [x] 路由按需引入 - [x] 支撑浏览器前进后退按钮的Tab管理组件 - [ ] 同路由多开Tab - [x] 国际化配置 - [ ] 完善说明文档 概述 ---- 基于`typescript`、`jsx`、`vue`、`ant-design-vue`的中后台模板 ![image](https://github.com/297854895/vue-tsx-admin/blob/master/home.png) ![image](https://github.com/297854895/vue-tsx-admin/blob/master/login.png) 下载与运行 ---- - 拉取项目 ```bash git clone https://github.com/297854895/vue-tsx-admin.git cd vue-tsx-admin ``` - 安装依赖 ``` yarn install ``` - 开发模式运行 ``` yarn run serve ``` - 编译项目 ``` yarn run build ``` - Lints and fixes files ``` yarn run lint ``` 环境和依赖 ---- - node - webpack - eslint - @vue/cli ~3 - [ant-design-vue](https://github.com/vueComponent/ant-design-vue) - Ant Design Of Vue 实现 目录结构 ---- + public color.less `动态主题色less文件` favicon.ico `系统图标` index.ejs `网页入口模板` + src + assets `资源存放` + components `组件存放` - Basic `系统基础组件` index.ts `抛出所有组件` + config `系统配置` cryptoKey.ts `本地记住密码加密的key(基于crypto-js的aes加密)` default.homeKey.ts `项目中默认的首页id` menu.ts `菜单的动态生成方法(基于routes.ts)` routes.ts `路由配置` routesInfo.ts `生成路由信息` systemLocalStoreKey.ts `系统基础组件需要缓存的数据key` themeColor.ts `主题色列表` + layouts `布局` - BasicLayout `系统基础布局(成功登录后)` - RouterLayout `系统多级路由容器` index.ts `抛出所有layouts` + locale `语言国际化配置` - en-US `english` - zh-CN `简体中文` - zh-TW `繁体中文` default.ts `初始化默认语言` index.ts `抛出所有语言` + localStore `路由页面数据持久化` + store `vuex状态集中管理` - models `分模块管理状态` actions.ts `根级action` clearStore.ts `初始化当前状态` extendsLocalStore.ts `路由数据持久化时,继承本地数据的方法` index.ts `生成状态库` muations `根级mutaion` types.ts `该目录下用到的类型声明约定` + style `样式` defaultThemeColor.ts `系统初始化时默认颜色` index.css `全局样式文件` reset.css `重置样式` theme.less `全局less变量定义` + utils `工具方法` appendLessFile.ts `添加主题色编译的color.less` createGuid.ts `创建不重复的guid` createThemeColorCss.ts `热编译主题色css` crypto.ts `本地记住密码加解密方法` enableRouterLocalStore.ts `开启路由数据持久化` event.ts `处理事件绑定` getClientHW.ts `获取客户端宽度与高度` getDeviceType.ts `获取客户端类型` validateLogin.ts `验证用户是否登录` + views `视图容器` - Login `登录界面` App.tsx `vue渲染根节点` main.ts `主入口` router.ts `路由生成` shims-other.d.ts `项目中需要添加的typescript配置额外声明` shims-tsx.d.ts `tsx类型的全局声明` shims-vue.d.ts `vue相关类型声明` + tests `单元测试` .eslintrc.js `eslint配置文件` babel.config.js `babel配置文件` tsconfig.json `typescript配置` vue.config.js `vue-cli配置重写` 约束 --- * 路由级组件容器置于`src/views`目录下 * 可复用功能性组件置于`src/components`目录下 * 系统配置文件置于`src/config`目录下 * 工具方法置放于`src/utils`目录下 系统配置 ---