repo_id stringlengths 6 101 | size int64 367 5.14M | file_path stringlengths 2 269 | content stringlengths 367 5.14M |
|---|---|---|---|
2977094657/DidaAPI | 7,837 | routers/tasks.py | """任务相关API路由"""
from fastapi import APIRouter, HTTPException, Query, Body
from typing import Optional
from models import ApiResponse
from services import dida_service
from utils import app_logger
router = APIRouter(prefix="/tasks", tags=["任务管理"])
@router.post("/set-auth",
response_model=ApiResponse,
summary="设置认证会话",
description="设置滴答清单API的认证令牌,用于后续API调用")
async def set_auth_session(
auth_token: str = Body(..., description="认证令牌(t cookie值)"),
csrf_token: str = Body(..., description="CSRF令牌")
) -> ApiResponse:
"""
设置认证会话
- **auth_token**: 认证令牌(从微信登录获得的 t cookie值)
- **csrf_token**: CSRF令牌(从微信登录获得的 _csrf_token值)
设置后可以调用其他需要认证的API接口
"""
try:
app_logger.info("设置认证会话")
if not auth_token or not csrf_token:
raise HTTPException(
status_code=400,
detail="认证令牌和CSRF令牌不能为空"
)
session_id = dida_service.set_auth_session(auth_token, csrf_token)
return ApiResponse(
code=200,
message="认证会话设置成功",
data={
"session_id": session_id,
"status": "已设置认证会话,可以调用其他API"
}
)
except HTTPException:
raise
except Exception as e:
app_logger.error(f"设置认证会话时发生错误: {e}")
raise HTTPException(
status_code=500,
detail=f"设置认证会话失败: {str(e)}"
)
@router.get("/all",
summary="获取所有任务",
description="获取当前用户的所有任务列表")
async def get_all_tasks():
"""
获取所有任务
返回当前用户的所有任务列表,包括:
- 任务ID、标题、内容
- 任务状态(0=未完成,2=已完成)
- 优先级、创建时间、修改时间
- 项目ID、标签等信息
**注意**: 需要先调用 `/tasks/set-auth` 设置认证会话
"""
try:
app_logger.info("请求获取所有任务")
result = await dida_service.get_all_tasks()
if not result:
return {"error": "获取任务失败,请稍后重试"}
# 记录日志
if 'error' in result:
app_logger.info(f"获取任务失败: {result.get('error')}")
else:
app_logger.info(f"任务获取完成")
# 直接返回原始响应
return result
except HTTPException:
raise
except Exception as e:
app_logger.error(f"获取任务时发生未知错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/summary",
response_model=ApiResponse,
summary="获取任务统计",
description="获取任务的统计信息")
async def get_tasks_summary() -> ApiResponse:
"""
获取任务统计
返回任务的统计信息:
- 总任务数
- 已完成任务数
- 未完成任务数
- 完成率等
"""
try:
app_logger.info("请求获取任务统计")
result = await dida_service.get_all_tasks()
if not result or "error" in result:
return {"error": "获取任务统计失败", "details": result}
# 统计任务信息 - 从原始响应中提取
total_tasks = 0
completed_tasks = 0
pending_tasks = 0
# 解析原始响应数据
if isinstance(result, dict) and 'syncTaskBean' in result:
task_data = result['syncTaskBean']
if 'update' in task_data:
raw_tasks = task_data['update']
total_tasks = len(raw_tasks)
for task in raw_tasks:
if task.get('status') == 2: # 已完成
completed_tasks += 1
else: # 未完成
pending_tasks += 1
completion_rate = (completed_tasks / total_tasks * 100) if total_tasks > 0 else 0
return ApiResponse(
code=200,
message="获取任务统计成功",
data={
"total_tasks": total_tasks,
"completed_tasks": completed_tasks,
"pending_tasks": pending_tasks,
"completion_rate": round(completion_rate, 2)
}
)
except HTTPException:
raise
except Exception as e:
app_logger.error(f"获取任务统计时发生错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/completed",
summary="获取已完成/已放弃任务",
description="获取已完成或已放弃的任务列表,支持分页获取")
async def get_completed_tasks(
to: Optional[str] = Query(None, description="分页参数,使用上次响应最后一个任务的completedTime字段,格式:2025-03-15T13:30:54.000+0000"),
status: str = Query("Completed", description="任务状态:Completed(已完成) 或 Abandoned(已放弃)")
):
"""
获取已完成/已放弃任务
支持分页获取已完成或已放弃的任务列表:
- **第一次请求**: 不传to参数,获取最新的任务
- **后续请求**: 使用上次响应最后一个任务的completedTime字段作为to参数
- **状态选择**: 通过status参数选择获取已完成(Completed)或已放弃(Abandoned)的任务
**分页机制说明**:
1. completedTime原始格式:2025-03-15T13:30:54.000+0000
2. to参数传入:使用completedTime的原始格式
3. API内部转换:2025-03-15T13:30:54.000+0000 → 2025-03-15 13:30:54
4. URL示例:https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Completed
**状态参数说明**:
- Completed: 获取已完成的任务
- Abandoned: 获取已放弃的任务
**注意**: 需要先完成微信登录获取认证会话
"""
try:
app_logger.info(f"请求获取任务,状态: {status},分页参数: {to}")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"}
# 调用服务获取任务
result = await dida_service.get_completed_tasks(to, status)
if not result:
return {"error": "service_error", "message": f"获取{status}任务失败,请稍后重试"}
# 记录日志
if 'error' in result:
app_logger.info(f"{status}任务获取失败: {result.get('error')}")
else:
task_count = len(result) if isinstance(result, list) else 0
app_logger.info(f"{status}任务获取完成,任务数: {task_count}")
# 如果有任务,记录最后一个任务的completedTime,便于下次分页
if isinstance(result, list) and len(result) > 0:
last_task = result[-1]
last_completed_time = last_task.get('completedTime')
if last_completed_time:
app_logger.info(f"最后一个任务的completedTime: {last_completed_time}")
# 直接返回原始响应
return result
except Exception as e:
app_logger.error(f"获取{status}任务时发生未知错误: {e}")
return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
@router.get("/trash",
summary="获取垃圾桶任务",
description="获取垃圾桶中的任务列表")
async def get_trash_tasks(
limit: int = Query(50, description="每页任务数量,默认50"),
task_type: int = Query(1, description="任务类型,默认1")
):
"""
获取垃圾桶任务
获取垃圾桶中的任务列表:
- **limit**: 每页返回的任务数量,默认50
- **task_type**: 任务类型,默认1
**响应格式**:
```json
{
"tasks": [...], // 任务列表
"next": 0 // 下一页标识
}
```
**注意**: 需要先完成微信登录获取认证会话
"""
try:
app_logger.info(f"请求获取垃圾桶任务,limit: {limit}, type: {task_type}")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"}
# 调用服务获取垃圾桶任务
result = await dida_service.get_trash_tasks(limit, task_type)
if not result:
return {"error": "service_error", "message": "获取垃圾桶任务失败,请稍后重试"}
# 记录日志
if 'error' in result:
app_logger.info(f"垃圾桶任务获取失败: {result.get('error')}")
else:
task_count = len(result.get('tasks', [])) if isinstance(result, dict) else 0
next_page = result.get('next', 0) if isinstance(result, dict) else 0
app_logger.info(f"垃圾桶任务获取完成,任务数: {task_count}, next: {next_page}")
# 直接返回原始响应
return result
except Exception as e:
app_logger.error(f"获取垃圾桶任务时发生未知错误: {e}")
return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
|
294coder/Efficient-MIF | 1,623 | configs/hyper_transformer_config.yaml | optimizer:
name: 'adam'
lr: !!float 1e-5
weight_decay: !!float 1e-6
lr_scheduler:
# name: 'multi_step'
# milestones: []
# gamma: 0.1
name: 'identity'
max_norm: null
path:
# train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5'
# val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5'
train_path: '/Data3/cao/cavex4/test_cavepatches128-2.h5'
# train_path: "/media/office-401/Elements SE/cao/cavex4/test_cavepatches64-2.h5"
# train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5"
val_path: "/Data3/cao/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5"
# train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/Chikusei_x4/train_Chikusei.h5"
# train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/Chikusei_x4/test_chikuseipatches340.h5"
# val_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/Chikusei_x4/test_Chikusei.h5"
# train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Pavia/Train_Pavia.h5"
# val_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Pavia/Test_Pavia.h5"
# train_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Botswana/Train_Botswana.h5"
# val_path: "/media/office-401/Elements SE/cao/ZiHanCao/datasets/HISI/data_Botswana/Test_Botswana.h5"
network_configs:
config: none
logger_config:
base_path: ./log_file/
name: hyper_transformer
file_mode: w |
294coder/Efficient-MIF | 1,412 | configs/hpmnet_config.yaml | optimizer:
name: "adamw"
lr: !!float 1e-4
weight_decay: !!float 1e-6
lr_scheduler:
name: "multi_step"
milestones: [1000, 1600]
gamma: 1
max_norm: 0.01
path:
# train_path: "/Data2/ZiHanCao/datasets/pansharpening/GF5_GF1/train_GF5_GF1.h5"
# val_path: "/Data2/Datasets/GF5_GF1/test_GF5_GF1-new.h5"
# train_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5"
# val_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5"
# train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5"
# val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5"
# train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5"
# train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"
# val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"
train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5"
val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5"
network_configs:
n_iter: 6
h_nc: 64
in_c: 9
out_c: 8
m_c: 3
nc: [80, 160, 320]
nb: 1
act_mode: "R"
downsample_mode: "strideconv"
upsample_mode: "convtranspose"
logger_config:
base_path: ./log_file/
name: hpmnet_gf2
file_mode: w
|
294coder/Efficient-MIF | 2,272 | configs/MGDN_config.yaml | optimizer:
name: "adamw"
lr: !!float 1e-3
weight_decay: !!float 1e-6
lr_scheduler:
name: 'cos_anneal_restart_reduce'
T_0: 70
T_mult: 2
lr_mult: 0.9
eta_min: !!float 8e-5
max_norm: 0.3
ema_decay: 0.99
path:
wv3_train_path: "/Data3/cao/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5"
wv3_val_path: "/Data3/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5"
gf2_train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5"
gf2_val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5"
qb_train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5"
qb_val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"
cave_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/train_cave(with_up)x4.h5"
cave_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5"
cave_x8_train_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/train_cave(with_up)x8_rgb_16.h5"
cave_x8_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/test_cave(with_up)x8_rgb.h5"
harvard_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5"
harvard_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/test_harvard(with_up)x4_rgb.h5"
harvard_x8_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/train_harvard(with_up)x8_rgb.h5"
harvard_x8_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/test_harvard(with_up)x8_rgb.h5"
msrs_base_dir: "/Data3/cao/ZiHanCao/datasets/MSRS"
llvip_base_dir: "/Data3/cao/ZiHanCao/datasets/LLVIP/data" # the mask is generated by using Ground-DINO-SAM project
tno_base_dir: "/Data3/cao/ZiHanCao/datasets/TNO"
roadscene_base_dir: "/Data3/cao/ZiHanCao/datasets/RoadSceneFusion"
roadscene_tno_joint_base_dir: "/Data3/cao/ZiHanCao/datasets/RoadScene_and_TNO"
med_harvard_base_dir: "/Data3/cao/ZiHanCao/datasets/MedHarvard"
network_configs:
GF_chans: 60
embed_dim: 60
depths:
- 6
- 6
num_heads:
- 6
- 6
window_size: 8
kernel_num:
- 1
- 1
- 1
mlp_ratio: 2.0
inference: false
logger_config:
base_path: ./log_file/
name: MGDN
file_mode: w
|
2977094657/DidaAPI | 9,036 | routers/export.py | """自定义导出功能API路由"""
from fastapi import APIRouter, HTTPException
from fastapi.responses import StreamingResponse
import io
import urllib.parse
from services.export_service import export_service
from services.dida_service import dida_service
from utils import app_logger
router = APIRouter(prefix="/custom", tags=["自定义接口"])
@router.get("/export/tasks/excel",
summary="导出任务到Excel",
description="导出所有任务到Excel文件,包含全部任务、已完成任务、放弃任务、垃圾桶任务四个工作表")
async def export_tasks_to_excel():
"""
导出任务到Excel
将用户的所有任务导出为Excel文件,包含以下工作表:
- **全部任务**: 当前所有任务(未完成和已完成)
- **已完成任务**: 历史已完成的任务
- **放弃任务**: 历史放弃的任务
- **垃圾桶任务**: 已删除的任务
每个工作表包含任务的完整字段信息,包括:
- 基本信息:任务ID、标题、内容、描述、项目信息
- 状态信息:任务状态、优先级、完成进度
- 时间信息:创建时间、修改时间、开始日期、截止日期
- 重复设置:重复标志、重复来源、首次重复日期
- 提醒设置:提醒配置、排除日期
- 层级关系:父任务、子任务关系
- 其他属性:标签、附件、评论数量等
**注意**: 需要先调用认证接口设置会话
"""
try:
app_logger.info("请求导出任务到Excel")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
raise HTTPException(
status_code=401,
detail="未设置认证会话,请先完成登录"
)
# 调用导出服务
result = await export_service.export_tasks_to_excel()
if 'error' in result:
app_logger.error(f"导出任务失败: {result['error']}")
raise HTTPException(
status_code=500,
detail=f"导出失败: {result['error']}"
)
app_logger.info(f"任务导出成功,文件大小: {result['size']} 字节")
# 对文件名进行URL编码以支持中文
encoded_filename = urllib.parse.quote(result['filename'], safe='')
# 返回文件下载响应
return StreamingResponse(
io.BytesIO(result['content']),
media_type=result['content_type'],
headers={
"Content-Disposition": f"attachment; filename*=UTF-8''{encoded_filename}"
}
)
except HTTPException:
raise
except Exception as e:
app_logger.error(f"导出任务时发生未知错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/export/tasks/excel/info",
summary="获取任务导出信息",
description="获取当前用户任务的统计信息,用于导出前预览")
async def get_export_info():
"""
获取任务导出信息
返回当前用户任务的统计信息,包括:
- 全部任务数量
- 已完成任务数量
- 垃圾桶任务数量
- 认证状态
**注意**: 需要先调用认证接口设置会话
"""
try:
app_logger.info("请求获取任务导出信息")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {
"error": "no_auth_session",
"message": "未设置认证会话,请先完成登录",
"auth_status": False
}
# 获取各类任务统计
stats = {
"auth_status": True,
"all_tasks_count": 0,
"completed_tasks_count": 0,
"abandoned_tasks_count": 0,
"trash_tasks_count": 0,
"session_info": session_status
}
# 获取全部任务统计
try:
all_tasks_result = await dida_service.get_all_tasks()
if all_tasks_result and 'error' not in all_tasks_result:
tasks = all_tasks_result.get('syncTaskBean', {}).get('update', [])
stats["all_tasks_count"] = len(tasks)
except Exception as e:
app_logger.warning(f"获取全部任务统计失败: {e}")
# 获取已完成任务统计(只获取第一页用于统计)
try:
completed_result = await dida_service.get_completed_tasks(None, "Completed")
if completed_result and 'error' not in completed_result:
if isinstance(completed_result, list):
stats["completed_tasks_count"] = len(completed_result)
except Exception as e:
app_logger.warning(f"获取已完成任务统计失败: {e}")
# 获取放弃任务统计
try:
abandoned_tasks_result = await dida_service.get_completed_tasks(None, "Abandoned")
if abandoned_tasks_result and 'error' not in abandoned_tasks_result:
if isinstance(abandoned_tasks_result, list):
stats["abandoned_tasks_count"] = len(abandoned_tasks_result)
except Exception as e:
app_logger.warning(f"获取放弃任务统计失败: {e}")
# 获取垃圾桶任务统计
try:
trash_result = await dida_service.get_trash_tasks()
if trash_result and 'error' not in trash_result:
tasks = trash_result.get('tasks', [])
stats["trash_tasks_count"] = len(tasks)
except Exception as e:
app_logger.warning(f"获取垃圾桶任务统计失败: {e}")
app_logger.info(f"任务统计获取完成: {stats}")
return stats
except Exception as e:
app_logger.error(f"获取任务导出信息时发生错误: {e}")
return {
"error": "server_error",
"message": f"服务器内部错误: {str(e)}",
"auth_status": False
}
@router.get("/export/focus/excel",
summary="导出专注记录到Excel",
description="导出所有专注记录到Excel文件,包含完整的专注时间线数据")
async def export_focus_records_to_excel():
"""
导出专注记录到Excel
将用户的所有专注记录导出为Excel文件,包含:
- **专注记录时间线**: 所有专注记录的详细信息
每个工作表包含专注记录的完整字段信息,包括:
- 基本信息:专注记录ID、开始时间、结束时间、创建时间
- 专注状态:专注状态、专注时长、暂停时长、实际专注时长
- 任务信息:任务ID、任务标题、项目ID、项目名称
- 标签信息:标签列表、标签ID列表
- 设备信息:设备类型、平台、应用版本
- 专注模式:专注模式、番茄钟时长、休息时长
- 其他属性:用户ID、时区、删除状态等
**注意**:
- 需要先调用认证接口设置会话
- 会自动分页获取所有历史专注记录
"""
try:
app_logger.info("请求导出专注记录到Excel")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
raise HTTPException(
status_code=401,
detail="未设置认证会话,请先完成登录"
)
# 调用导出服务
result = await export_service.export_focus_records_to_excel()
if 'error' in result:
app_logger.error(f"导出专注记录失败: {result['error']}")
raise HTTPException(
status_code=500,
detail=f"导出失败: {result['error']}"
)
app_logger.info(f"专注记录导出成功,文件大小: {result['size']} 字节")
# 对文件名进行URL编码以支持中文
encoded_filename = urllib.parse.quote(result['filename'], safe='')
# 返回文件下载响应
return StreamingResponse(
io.BytesIO(result['content']),
media_type=result['content_type'],
headers={
"Content-Disposition": f"attachment; filename*=UTF-8''{encoded_filename}"
}
)
except HTTPException:
raise
except Exception as e:
app_logger.error(f"导出专注记录时发生未知错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/export/focus/excel/info",
summary="获取专注记录导出信息",
description="获取当前用户专注记录的统计信息,用于导出前预览")
async def get_focus_export_info():
"""
获取专注记录导出信息
返回当前用户专注记录的统计信息,包括:
- 专注记录总数量(预估)
- 认证状态
**注意**: 需要先调用认证接口设置会话
"""
try:
app_logger.info("请求获取专注记录导出信息")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {
"error": "no_auth_session",
"message": "未设置认证会话,请先完成登录",
"auth_status": False
}
# 获取专注记录统计(只获取第一页用于预估)
stats = {
"auth_status": True,
"focus_records_count_estimate": 0,
"session_info": session_status,
"note": "专注记录数量为预估值,实际导出时会获取所有历史数据"
}
try:
# 获取认证信息
current_session = dida_service.current_session
auth_token = current_session['auth_token']
csrf_token = current_session['csrf_token']
# 获取第一页专注记录用于预估
from services.pomodoro_service import pomodoro_service
result = await pomodoro_service.get_focus_timeline(auth_token, csrf_token, None)
if result and 'error' not in result:
if isinstance(result, list):
# 如果第一页有31条记录,预估可能有更多数据
first_page_count = len(result)
if first_page_count >= 31:
stats["focus_records_count_estimate"] = f"{first_page_count}+ (需要分页获取完整数据)"
else:
stats["focus_records_count_estimate"] = first_page_count
else:
stats["focus_records_count_estimate"] = "无法预估"
except Exception as e:
app_logger.warning(f"获取专注记录统计失败: {e}")
stats["focus_records_count_estimate"] = "获取失败"
app_logger.info(f"专注记录统计获取完成: {stats}")
return stats
except Exception as e:
app_logger.error(f"获取专注记录导出信息时发生错误: {e}")
return {
"error": "server_error",
"message": f"服务器内部错误: {str(e)}",
"auth_status": False
}
|
2977094657/DidaAPI | 5,669 | routers/system.py | """系统相关API路由"""
from fastapi import APIRouter
from typing import Dict, Any
from core import urls
from models import ApiResponse
from utils import app_logger
router = APIRouter(prefix="/system", tags=["系统管理"])
@router.get("/urls",
response_model=ApiResponse,
summary="获取所有URL配置",
description="获取项目中管理的所有URL和外部链接配置")
async def get_all_urls() -> ApiResponse:
"""
获取所有URL配置
返回项目中统一管理的所有URL和外部链接
"""
try:
all_urls = urls.get_all_external_urls()
api_endpoints = urls.get_api_endpoints()
return ApiResponse(
code=200,
message="获取URL配置成功",
data={
"external_urls": all_urls,
"api_endpoints": api_endpoints
}
)
except Exception as e:
app_logger.error(f"获取URL配置时发生错误: {e}")
return ApiResponse(
code=500,
message=f"获取URL配置失败: {str(e)}",
data=None
)
@router.get("/urls/wechat",
response_model=ApiResponse,
summary="获取微信相关URL",
description="获取微信开放平台相关的所有URL配置")
async def get_wechat_urls() -> ApiResponse:
"""
获取微信相关URL
返回微信开放平台相关的URL配置
"""
try:
return ApiResponse(
code=200,
message="获取微信URL配置成功",
data={
"wechat_urls": urls.WECHAT_URLS,
"wechat_config": urls.WECHAT_CONFIG,
"helper_functions": [
"build_wechat_qr_url(state)",
"build_wechat_poll_url(uuid, timestamp)",
"build_wechat_validate_url(code, state)"
]
}
)
except Exception as e:
app_logger.error(f"获取微信URL配置时发生错误: {e}")
return ApiResponse(
code=500,
message=f"获取微信URL配置失败: {str(e)}",
data=None
)
@router.get("/urls/dida",
response_model=ApiResponse,
summary="获取滴答清单API URL",
description="获取滴答清单API相关的所有URL配置")
async def get_dida_urls() -> ApiResponse:
"""
获取滴答清单API URL
返回滴答清单API相关的URL配置
"""
try:
return ApiResponse(
code=200,
message="获取滴答清单URL配置成功",
data={
"base_config": urls.DIDA_API_BASE,
"auth_apis": urls.DIDA_AUTH_APIS,
"task_apis": urls.DIDA_TASK_APIS,
"project_apis": urls.DIDA_PROJECT_APIS,
"statistics_apis": urls.DIDA_STATISTICS_APIS,
"pomodoro_apis": urls.DIDA_POMODORO_APIS,
"habit_apis": urls.DIDA_HABIT_APIS,
"helper_functions": [
"build_dida_api_url(endpoint)"
]
}
)
except Exception as e:
app_logger.error(f"获取滴答清单URL配置时发生错误: {e}")
return ApiResponse(
code=500,
message=f"获取滴答清单URL配置失败: {str(e)}",
data=None
)
@router.get("/urls/docs",
response_model=ApiResponse,
summary="获取文档链接",
description="获取官方文档和技术参考链接")
async def get_doc_urls() -> ApiResponse:
"""
获取文档链接
返回官方文档和技术参考链接
"""
try:
return ApiResponse(
code=200,
message="获取文档链接成功",
data={
"official_docs": urls.OFFICIAL_DOCS,
"tech_references": urls.TECH_REFERENCES,
"project_docs": {
"frontend_docs": "http://localhost:3000",
"api_docs": "http://localhost:3000/api/",
"url_management": "http://localhost:3000/api/url-management",
"wechat_flow": "http://localhost:3000/api/auth/wechat-login-flow"
}
}
)
except Exception as e:
app_logger.error(f"获取文档链接时发生错误: {e}")
return ApiResponse(
code=500,
message=f"获取文档链接失败: {str(e)}",
data=None
)
@router.get("/info",
response_model=ApiResponse,
summary="获取系统信息",
description="获取项目的基本信息和配置概览")
async def get_system_info() -> ApiResponse:
"""
获取系统信息
返回项目的基本信息和配置概览
"""
try:
from core import config
return ApiResponse(
code=200,
message="获取系统信息成功",
data={
"project_name": "滴答清单Web端API接口",
"version": "1.0.0",
"description": "滴答清单原始API的封装层,提供简化的接口和完整的API文档",
"features": [
"微信扫码登录",
"任务管理",
"会话持久化",
"统一URL管理",
"完整API文档"
],
"url_management": {
"total_external_urls": len(urls.WECHAT_URLS) + len(urls.DIDA_API_BASE) + len(urls.OFFICIAL_DOCS) + len(urls.TECH_REFERENCES),
"total_api_endpoints": len(urls.DIDA_AUTH_APIS) + len(urls.DIDA_TASK_APIS) + len(urls.DIDA_PROJECT_APIS) + len(urls.DIDA_STATISTICS_APIS) + len(urls.DIDA_POMODORO_APIS) + len(urls.DIDA_HABIT_APIS),
"management_file": "core/urls.py"
},
"config": {
"app": config.app,
"request_config": config.get('request_config', {}),
"database": config.database
}
}
)
except Exception as e:
app_logger.error(f"获取系统信息时发生错误: {e}")
return ApiResponse(
code=500,
message=f"获取系统信息失败: {str(e)}",
data=None
)
|
281677160/openwrt-package | 66,136 | luci-app-passwall/root/usr/share/passwall/subscribe.lua | #!/usr/bin/lua
------------------------------------------------
-- @author William Chan <root@williamchan.me>
------------------------------------------------
require 'luci.util'
require 'luci.jsonc'
require 'luci.sys'
local appname = 'passwall'
local api = require ("luci.passwall.api")
local datatypes = require "luci.cbi.datatypes"
-- these global functions are accessed all the time by the event handler
-- so caching them is worth the effort
local tinsert = table.insert
local ssub, slen, schar, sbyte, sformat, sgsub = string.sub, string.len, string.char, string.byte, string.format, string.gsub
local split = api.split
local jsonParse, jsonStringify = luci.jsonc.parse, luci.jsonc.stringify
local base64Decode = api.base64Decode
local uci = api.uci
local fs = api.fs
uci:revert(appname)
local has_ss = api.is_finded("ss-redir")
local has_ss_rust = api.is_finded("sslocal")
local has_ssr = api.is_finded("ssr-local") and api.is_finded("ssr-redir")
local has_trojan_plus = api.is_finded("trojan-plus")
local has_singbox = api.finded_com("sing-box")
local has_xray = api.finded_com("xray")
local has_hysteria2 = api.finded_com("hysteria")
local allowInsecure_default = nil
-- 取节点使用core类型(节点订阅页面未设置时,自动取默认)
local function get_core(field, candidates)
local v = uci:get(appname, "@global_subscribe[0]", field)
if not v or v == "" then
for _, c in ipairs(candidates) do
if c[1] then return c[2] end
end
end
return v
end
local ss_type_default = get_core("ss_type", {{has_ss,"shadowsocks-libev"},{has_ss_rust,"shadowsocks-rust"},{has_singbox,"sing-box"},{has_xray,"xray"}})
local trojan_type_default = get_core("trojan_type", {{has_trojan_plus,"trojan-plus"},{has_singbox,"sing-box"},{has_xray,"xray"}})
local vmess_type_default = get_core("vmess_type", {{has_xray,"xray"},{has_singbox,"sing-box"}})
local vless_type_default = get_core("vless_type", {{has_xray,"xray"},{has_singbox,"sing-box"}})
local hysteria2_type_default = get_core("hysteria2_type", {{has_hysteria2,"hysteria2"},{has_singbox,"sing-box"}})
----
local domain_strategy_default = uci:get(appname, "@global_subscribe[0]", "domain_strategy") or ""
local domain_strategy_node = ""
local preproxy_node_group, to_node_group, chain_node_type = "", "", ""
-- 判断是否过滤节点关键字
local filter_keyword_mode_default = uci:get(appname, "@global_subscribe[0]", "filter_keyword_mode") or "0"
local filter_keyword_discard_list_default = uci:get(appname, "@global_subscribe[0]", "filter_discard_list") or {}
local filter_keyword_keep_list_default = uci:get(appname, "@global_subscribe[0]", "filter_keep_list") or {}
local function is_filter_keyword(value)
if filter_keyword_mode_default == "1" then
for k,v in ipairs(filter_keyword_discard_list_default) do
if value:find(v, 1, true) then
return true
end
end
elseif filter_keyword_mode_default == "2" then
local result = true
for k,v in ipairs(filter_keyword_keep_list_default) do
if value:find(v, 1, true) then
result = false
end
end
return result
elseif filter_keyword_mode_default == "3" then
local result = false
for k,v in ipairs(filter_keyword_discard_list_default) do
if value:find(v, 1, true) then
result = true
end
end
for k,v in ipairs(filter_keyword_keep_list_default) do
if value:find(v, 1, true) then
result = false
end
end
return result
elseif filter_keyword_mode_default == "4" then
local result = true
for k,v in ipairs(filter_keyword_keep_list_default) do
if value:find(v, 1, true) then
result = false
end
end
for k,v in ipairs(filter_keyword_discard_list_default) do
if value:find(v, 1, true) then
result = true
end
end
return result
end
return false
end
local nodeResult = {} -- update result
local isDebug = false
local log = function(...)
if isDebug == true then
local result = os.date("%Y-%m-%d %H:%M:%S: ") .. table.concat({...}, " ")
print(result)
else
api.log(...)
end
end
local nodes_table = {}
for k, e in ipairs(api.get_valid_nodes()) do
if e.node_type == "normal" then
nodes_table[#nodes_table + 1] = e
end
end
-- 获取各项动态配置的当前服务器,可以用 get 和 set, get必须要获取到节点表
local CONFIG = {}
do
local function import_config(protocol)
local name = string.upper(protocol)
local szType = "@global[0]"
local option = protocol .. "_node"
local node_id = uci:get(appname, szType, option)
CONFIG[#CONFIG + 1] = {
log = true,
remarks = name .. "节点",
currentNode = node_id and uci:get_all(appname, node_id) or nil,
set = function(o, server)
uci:set(appname, szType, option, server)
o.newNodeId = server
end
}
end
import_config("tcp")
import_config("udp")
if true then
local i = 0
local option = "node"
uci:foreach(appname, "socks", function(t)
i = i + 1
local id = t[".name"]
local node_id = t[option]
CONFIG[#CONFIG + 1] = {
log = true,
id = id,
remarks = "Socks节点列表[" .. i .. "]",
currentNode = node_id and uci:get_all(appname, node_id) or nil,
set = function(o, server)
if not server or server == "" then
if #nodes_table > 0 then
server = nodes_table[1][".name"]
end
end
uci:set(appname, t[".name"], option, server)
o.newNodeId = server
end
}
if t.autoswitch_backup_node and #t.autoswitch_backup_node > 0 then
local flag = "Socks节点列表[" .. i .. "]备用节点的列表"
local currentNodes = {}
local newNodes = {}
for k, node_id in ipairs(t.autoswitch_backup_node) do
if node_id then
local currentNode = uci:get_all(appname, node_id) or nil
if currentNode then
currentNodes[#currentNodes + 1] = {
log = true,
remarks = flag .. "[" .. k .. "]",
currentNode = currentNode,
set = function(o, server)
if server and server ~= "nil" then
table.insert(o.newNodes, server)
end
end
}
end
end
end
CONFIG[#CONFIG + 1] = {
remarks = flag,
currentNodes = currentNodes,
newNodes = newNodes,
set = function(o, newNodes)
if o then
if not newNodes then newNodes = o.newNodes end
uci:set_list(appname, id, "autoswitch_backup_node", newNodes or {})
end
end
}
end
end)
end
if true then
local i = 0
local option = "lbss"
local function is_ip_port(str)
if type(str) ~= "string" then return false end
local ip, port = str:match("^([%d%.]+):(%d+)$")
return ip and datatypes.ipaddr(ip) and tonumber(port) and tonumber(port) <= 65535
end
uci:foreach(appname, "haproxy_config", function(t)
i = i + 1
local node_id = t[option]
CONFIG[#CONFIG + 1] = {
log = true,
id = t[".name"],
remarks = "HAProxy负载均衡节点列表[" .. i .. "]",
currentNode = node_id and uci:get_all(appname, node_id) or nil,
set = function(o, server)
-- 如果当前 lbss 值不是 ip:port 格式,才进行修改
if not is_ip_port(t[option]) then
uci:set(appname, t[".name"], option, server)
o.newNodeId = server
end
end,
delete = function(o)
-- 如果当前 lbss 值不是 ip:port 格式,才进行删除
if not is_ip_port(t[option]) then
uci:delete(appname, t[".name"])
end
end
}
end)
end
if true then
local i = 0
local options = {"tcp", "udp"}
uci:foreach(appname, "acl_rule", function(t)
i = i + 1
for index, value in ipairs(options) do
local option = value .. "_node"
local node_id = t[option]
CONFIG[#CONFIG + 1] = {
log = true,
id = t[".name"],
remarks = "访问控制列表[" .. i .. "]",
currentNode = node_id and uci:get_all(appname, node_id) or nil,
set = function(o, server)
uci:set(appname, t[".name"], option, server)
o.newNodeId = server
end
}
end
end)
end
uci:foreach(appname, "nodes", function(node)
local node_id = node[".name"]
if node.protocol and node.protocol == '_shunt' then
local rules = {}
uci:foreach(appname, "shunt_rules", function(e)
if e[".name"] and e.remarks then
table.insert(rules, e)
end
end)
table.insert(rules, {
[".name"] = "default_node",
remarks = "默认"
})
table.insert(rules, {
[".name"] = "main_node",
remarks = "默认前置"
})
for k, e in pairs(rules) do
local _node_id = node[e[".name"]] or nil
if _node_id and api.parseURL(_node_id) then
else
CONFIG[#CONFIG + 1] = {
log = false,
currentNode = _node_id and uci:get_all(appname, _node_id) or nil,
remarks = "分流" .. e.remarks .. "节点",
set = function(o, server)
if not server then server = "" end
uci:set(appname, node_id, e[".name"], server)
o.newNodeId = server
end
}
end
end
elseif node.protocol and node.protocol == '_balancing' then
local flag = "Xray负载均衡节点[" .. node_id .. "]列表"
local currentNodes = {}
local newNodes = {}
if node.balancing_node then
for k, node in pairs(node.balancing_node) do
currentNodes[#currentNodes + 1] = {
log = true,
node = node,
currentNode = node and uci:get_all(appname, node) or nil,
remarks = node,
set = function(o, server)
if o and server and server ~= "nil" then
table.insert(o.newNodes, server)
end
end
}
end
end
CONFIG[#CONFIG + 1] = {
remarks = flag,
currentNodes = currentNodes,
newNodes = newNodes,
set = function(o, newNodes)
if o then
if not newNodes then newNodes = o.newNodes end
uci:set_list(appname, node_id, "balancing_node", newNodes or {})
end
end
}
--后备节点
local currentNode = uci:get_all(appname, node_id) or nil
if currentNode and currentNode.fallback_node then
CONFIG[#CONFIG + 1] = {
log = true,
id = node_id,
remarks = "Xray负载均衡节点[" .. node_id .. "]后备节点",
currentNode = uci:get_all(appname, currentNode.fallback_node) or nil,
set = function(o, server)
uci:set(appname, node_id, "fallback_node", server)
o.newNodeId = server
end,
delete = function(o)
uci:delete(appname, node_id, "fallback_node")
end
}
end
elseif node.protocol and node.protocol == '_urltest' then
local flag = "Sing-Box URLTest节点[" .. node_id .. "]列表"
local currentNodes = {}
local newNodes = {}
if node.urltest_node then
for k, node in pairs(node.urltest_node) do
currentNodes[#currentNodes + 1] = {
log = true,
node = node,
currentNode = node and uci:get_all(appname, node) or nil,
remarks = node,
set = function(o, server)
if o and server and server ~= "nil" then
table.insert(o.newNodes, server)
end
end
}
end
end
CONFIG[#CONFIG + 1] = {
remarks = flag,
currentNodes = currentNodes,
newNodes = newNodes,
set = function(o, newNodes)
if o then
if not newNodes then newNodes = o.newNodes end
uci:set_list(appname, node_id, "urltest_node", newNodes or {})
end
end
}
else
--前置代理节点
local currentNode = uci:get_all(appname, node_id) or nil
if currentNode and currentNode.preproxy_node then
CONFIG[#CONFIG + 1] = {
log = true,
id = node_id,
remarks = "节点[" .. node_id .. "]前置代理节点",
currentNode = uci:get_all(appname, currentNode.preproxy_node) or nil,
set = function(o, server)
uci:set(appname, node_id, "preproxy_node", server)
o.newNodeId = server
end,
delete = function(o)
uci:delete(appname, node_id, "preproxy_node")
end
}
end
--落地节点
local currentNode = uci:get_all(appname, node_id) or nil
if currentNode and currentNode.to_node then
CONFIG[#CONFIG + 1] = {
log = true,
id = node_id,
remarks = "节点[" .. node_id .. "]落地节点",
currentNode = uci:get_all(appname, currentNode.to_node) or nil,
set = function(o, server)
uci:set(appname, node_id, "to_node", server)
o.newNodeId = server
end,
delete = function(o)
uci:delete(appname, node_id, "to_node")
end
}
end
end
end)
for k, v in pairs(CONFIG) do
if v.currentNodes and type(v.currentNodes) == "table" then
for kk, vv in pairs(v.currentNodes) do
if vv.currentNode == nil then
CONFIG[k].currentNodes[kk] = nil
end
end
else
if v.currentNode == nil then
if v.delete then
v.delete()
end
CONFIG[k] = nil
end
end
end
end
local function UrlEncode(szText)
return szText:gsub("([^%w%-_%.%~])", function(c)
return string.format("%%%02X", string.byte(c))
end)
end
local function UrlDecode(szText)
return szText and szText:gsub("+", " "):gsub("%%(%x%x)", function(h)
return string.char(tonumber(h, 16))
end) or nil
end
-- 取机场信息(剩余流量、到期时间)
local subscribe_info = {}
local function get_subscribe_info(cfgid, value)
if type(cfgid) ~= "string" or cfgid == "" or type(value) ~= "string" then
return
end
value = value:gsub("%s+", "")
local expired_date = value:match("套餐到期:(.+)")
local rem_traffic = value:match("剩余流量:(.+)")
subscribe_info[cfgid] = subscribe_info[cfgid] or {expired_date = "", rem_traffic = ""}
if expired_date then
subscribe_info[cfgid]["expired_date"] = expired_date
end
if rem_traffic then
subscribe_info[cfgid]["rem_traffic"] = rem_traffic
end
end
-- 处理数据
local function processData(szType, content, add_mode, add_from)
--log(content, add_mode, add_from)
local result = {
timeout = 60,
add_mode = add_mode, --0为手动配置,1为导入,2为订阅
add_from = add_from
}
--ssr://base64(host:port:protocol:method:obfs:base64pass/?obfsparam=base64param&protoparam=base64param&remarks=base64remarks&group=base64group&udpport=0&uot=0)
if szType == 'ssr' then
if not has_ssr then
log("跳过 SSR 节点,因未安装 SSR 核心程序 shadowsocksr-libev。")
return nil
end
result.type = "SSR"
local dat = split(content, "/%?")
local hostInfo = split(dat[1], ':')
if dat[1]:match('%[(.*)%]') then
result.address = dat[1]:match('%[(.*)%]')
else
result.address = hostInfo[#hostInfo-5]
end
result.port = hostInfo[#hostInfo-4]
result.protocol = hostInfo[#hostInfo-3]
result.method = hostInfo[#hostInfo-2]
result.obfs = hostInfo[#hostInfo-1]
result.password = base64Decode(hostInfo[#hostInfo])
local params = {}
for _, v in pairs(split(dat[2], '&')) do
local t = split(v, '=')
params[t[1]] = t[2]
end
result.obfs_param = base64Decode(params.obfsparam)
result.protocol_param = base64Decode(params.protoparam)
local group = base64Decode(params.group)
if group then result.group = group end
result.remarks = base64Decode(params.remarks)
elseif szType == 'vmess' then
local info = jsonParse(content)
if vmess_type_default == "sing-box" and has_singbox then
result.type = 'sing-box'
elseif vmess_type_default == "xray" and has_xray then
result.type = "Xray"
else
log("跳过 VMess 节点,因未适配到 VMess 核心程序,或未正确设置节点使用类型。")
return nil
end
result.alter_id = info.aid
result.address = info.add
result.port = info.port
result.protocol = 'vmess'
result.uuid = info.id
result.remarks = info.ps
-- result.mux = 1
-- result.mux_concurrency = 8
if not info.net then info.net = "tcp" end
info.net = string.lower(info.net)
if result.type == "sing-box" and info.net == "raw" then
info.net = "tcp"
elseif result.type == "Xray" and info.net == "tcp" then
info.net = "raw"
end
if info.net == 'h2' or info.net == 'http' then
info.net = "http"
result.transport = (result.type == "Xray") and "xhttp" or "http"
else
result.transport = info.net
end
if info.net == 'ws' then
result.ws_host = info.host
result.ws_path = info.path
if result.type == "sing-box" and info.path then
local ws_path_dat = split(info.path, "?")
local ws_path = ws_path_dat[1]
local ws_path_params = {}
for _, v in pairs(split(ws_path_dat[2], '&')) do
local t = split(v, '=')
ws_path_params[t[1]] = t[2]
end
if ws_path_params.ed and tonumber(ws_path_params.ed) then
result.ws_path = ws_path
result.ws_enableEarlyData = "1"
result.ws_maxEarlyData = tonumber(ws_path_params.ed)
result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol"
end
end
end
if info.net == "http" then
if result.type == "Xray" then
result.xhttp_mode = "stream-one"
result.xhttp_host = info.host
result.xhttp_path = info.path
else
result.http_host = (info.host and info.host ~= "") and { info.host } or nil
result.http_path = info.path
end
end
if info.net == 'raw' or info.net == 'tcp' then
if info.type and info.type ~= "http" then
info.type = "none"
end
result.tcp_guise = info.type
result.tcp_guise_http_host = (info.host and info.host ~= "") and { info.host } or nil
result.tcp_guise_http_path = (info.path and info.path ~= "") and { info.path } or nil
end
if info.net == 'kcp' or info.net == 'mkcp' then
info.net = "mkcp"
result.mkcp_guise = info.type
result.mkcp_mtu = 1350
result.mkcp_tti = 50
result.mkcp_uplinkCapacity = 5
result.mkcp_downlinkCapacity = 20
result.mkcp_readBufferSize = 2
result.mkcp_writeBufferSize = 2
result.mkcp_seed = info.seed
end
if info.net == 'quic' then
result.quic_guise = info.type
result.quic_key = info.key
result.quic_security = info.securty
end
if info.net == 'grpc' then
result.grpc_serviceName = info.path
end
if info.net == 'xhttp' then
result.xhttp_host = info.host
result.xhttp_path = info.path
end
if info.net == 'httpupgrade' then
result.httpupgrade_host = info.host
result.httpupgrade_path = info.path
end
if not info.security then result.security = "auto" end
if info.tls == "tls" or info.tls == "1" then
result.tls = "1"
result.tls_serverName = (info.sni and info.sni ~= "") and info.sni or info.host
info.allowinsecure = info.allowinsecure or info.insecure
if info.allowinsecure and (info.allowinsecure == "1" or info.allowinsecure == "0") then
result.tls_allowInsecure = info.allowinsecure
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
else
result.tls = "0"
end
if result.type == "sing-box" and (result.transport == "mkcp" or result.transport == "xhttp") then
log("跳过节点:" .. result.remarks ..",因Sing-Box不支持" .. szType .. "协议的" .. result.transport .. "传输方式,需更换Xray。")
return nil
end
elseif szType == "ss" then
if ss_type_default == "shadowsocks-libev" and has_ss then
result.type = "SS"
elseif ss_type_default == "shadowsocks-rust" and has_ss_rust then
result.type = 'SS-Rust'
elseif ss_type_default == "xray" and has_xray then
result.type = 'Xray'
result.protocol = 'shadowsocks'
result.transport = 'raw'
elseif ss_type_default == "sing-box" and has_singbox then
result.type = 'sing-box'
result.protocol = 'shadowsocks'
else
log("跳过 SS 节点,因未适配到 SS 核心程序,或未正确设置节点使用类型。")
return nil
end
--SS-URI = "ss://" userinfo "@" hostname ":" port [ "/" ] [ "?" plugin ] [ "#" tag ]
--userinfo = websafe-base64-encode-utf8(method ":" password)
--ss://YWVzLTEyOC1nY206dGVzdA@192.168.100.1:8888#Example1
--ss://cmM0LW1kNTpwYXNzd2Q@192.168.100.1:8888/?plugin=obfs-local%3Bobfs%3Dhttp#Example2
--ss://2022-blake3-aes-256-gcm:YctPZ6U7xPPcU%2Bgp3u%2B0tx%2FtRizJN9K8y%2BuKlW2qjlI%3D@192.168.100.1:8888#Example3
--ss://2022-blake3-aes-256-gcm:YctPZ6U7xPPcU%2Bgp3u%2B0tx%2FtRizJN9K8y%2BuKlW2qjlI%3D@192.168.100.1:8888/?plugin=v2ray-plugin%3Bserver#Example3
--ss://Y2hhY2hhMjAtaWV0Zi1wb2x5MTMwNTp0ZXN0@xxxxxx.com:443?type=ws&path=%2Ftestpath&host=xxxxxx.com&security=tls&fp=&alpn=h3%2Ch2%2Chttp%2F1.1&sni=xxxxxx.com#test-1%40ss
--ss://Y2hhY2hhMjAtaWV0Zi1wb2x5MTMwNTp4eHh4eHhAeHh4eC54eHh4eC5jb206NTYwMDE#Hong%20Kong-01
local idx_sp = content:find("#") or 0
local alias = ""
if idx_sp > 0 then
alias = content:sub(idx_sp + 1, -1)
end
result.remarks = UrlDecode(alias)
local info = content:sub(1, idx_sp - 1):gsub("/%?", "?")
local params = {}
if info:find("%?") then
local find_index = info:find("%?")
local query = split(info, "%?")
for _, v in pairs(split(query[2], '&')) do
local t = split(v, '=')
if #t >= 2 then params[t[1]] = UrlDecode(t[2]) end
end
if params.plugin then
local plugin_info = params.plugin
local idx_pn = plugin_info:find(";")
if idx_pn then
result.plugin = plugin_info:sub(1, idx_pn - 1)
result.plugin_opts = plugin_info:sub(idx_pn + 1, #plugin_info)
else
result.plugin = plugin_info
end
end
if result.plugin and result.plugin == "simple-obfs" then
result.plugin = "obfs-local"
end
info = info:sub(1, find_index - 1)
end
local hostInfo = split(base64Decode(UrlDecode(info)), "@")
if hostInfo and #hostInfo > 0 then
local host_port = hostInfo[#hostInfo]
-- [2001:4860:4860::8888]:443
-- 8.8.8.8:443
if host_port:find(":") then
local sp = split(host_port, ":")
result.port = sp[#sp]
if api.is_ipv6addrport(host_port) then
result.address = api.get_ipv6_only(host_port)
else
result.address = sp[1]
end
else
result.address = host_port
end
local userinfo = nil
if #hostInfo > 2 then
userinfo = {}
for i = 1, #hostInfo - 1 do
tinsert(userinfo, hostInfo[i])
end
userinfo = table.concat(userinfo, '@')
else
userinfo = base64Decode(hostInfo[1])
end
local method, password
if userinfo:find(":") then
method = userinfo:sub(1, userinfo:find(":") - 1)
password = userinfo:sub(userinfo:find(":") + 1, #userinfo)
else
password = hostInfo[1] --一些链接用明文uuid做密码
end
-- 判断密码是否经过url编码
local function isURLEncodedPassword(pwd)
if not pwd:find("%%[0-9A-Fa-f][0-9A-Fa-f]") then
return false
end
local ok, decoded = pcall(UrlDecode, pwd)
return ok and UrlEncode(decoded) == pwd
end
local decoded = UrlDecode(password)
if isURLEncodedPassword(password) and decoded then
password = decoded
end
local _method = (method or "none"):lower()
method = (_method == "chacha20-poly1305" and "chacha20-ietf-poly1305") or
(_method == "xchacha20-poly1305" and "xchacha20-ietf-poly1305") or _method
result.method = method
result.password = password
if has_xray and (result.type ~= 'Xray' and result.type ~= 'sing-box' and params.type) then
result.type = 'Xray'
result.protocol = 'shadowsocks'
elseif has_singbox and (result.type ~= 'Xray' and result.type ~= 'sing-box' and params.type) then
result.type = 'sing-box'
result.protocol = 'shadowsocks'
end
if result.plugin then
if result.type == 'Xray' then
-- obfs-local插件转换成xray支持的格式
if result.plugin ~= "obfs-local" then
result.error_msg = "Xray不支持 " .. result.plugin .. " 插件."
else
local obfs = result.plugin_opts:match("obfs=([^;]+)") or ""
local obfs_host = result.plugin_opts:match("obfs%-host=([^;]+)") or ""
if obfs == "" or obfs_host == "" then
result.error_msg = "SS " .. result.plugin .. " 插件选项不完整."
end
if obfs == "http" then
result.transport = "raw"
result.tcp_guise = "http"
result.tcp_guise_http_host = (obfs_host and obfs_host ~= "") and { obfs_host } or nil
result.tcp_guise_http_path = { "/" }
elseif obfs == "tls" then
result.tls = "1"
result.tls_serverName = obfs_host
result.tls_allowInsecure = "1"
end
result.plugin = nil
result.plugin_opts = nil
end
else
result.plugin_enabled = "1"
end
end
if result.type == "SS" then
local aead2022_methods = { "2022-blake3-aes-128-gcm", "2022-blake3-aes-256-gcm", "2022-blake3-chacha20-poly1305" }
local aead2022 = false
for k, v in ipairs(aead2022_methods) do
if method:lower() == v:lower() then
aead2022 = true
end
end
if aead2022 then
-- shadowsocks-libev 不支持2022加密
result.error_msg = "shadowsocks-libev 不支持2022加密."
end
end
if params.type then
params.type = string.lower(params.type)
if result.type == "sing-box" and params.type == "raw" then
params.type = "tcp"
elseif result.type == "Xray" and params.type == "tcp" then
params.type = "raw"
end
if params.type == "h2" or params.type == "http" then
params.type = "http"
result.transport = (result.type == "Xray") and "xhttp" or "http"
else
result.transport = params.type
end
if result.type ~= "SS-Rust" and result.type ~= "SS" then
if params.type == 'ws' then
result.ws_host = params.host
result.ws_path = params.path
if result.type == "sing-box" and params.path then
local ws_path_dat = split(params.path, "%?")
local ws_path = ws_path_dat[1]
local ws_path_params = {}
for _, v in pairs(split(ws_path_dat[2], '&')) do
local t = split(v, '=')
ws_path_params[t[1]] = t[2]
end
if ws_path_params.ed and tonumber(ws_path_params.ed) then
result.ws_path = ws_path
result.ws_enableEarlyData = "1"
result.ws_maxEarlyData = tonumber(ws_path_params.ed)
result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol"
end
end
end
if params.type == "http" then
if result.type == "sing-box" then
result.transport = "http"
result.http_host = (params.host and params.host ~= "") and { params.host } or nil
result.http_path = params.path
elseif result.type == "Xray" then
result.transport = "xhttp"
result.xhttp_mode = "stream-one"
result.xhttp_host = params.host
result.xhttp_path = params.path
end
end
if params.type == 'raw' or params.type == 'tcp' then
result.tcp_guise = params.headerType or "none"
result.tcp_guise_http_host = (params.host and params.host ~= "") and { params.host } or nil
result.tcp_guise_http_path = (params.path and params.path ~= "") and { params.path } or nil
end
if params.type == 'kcp' or params.type == 'mkcp' then
result.transport = "mkcp"
result.mkcp_guise = params.headerType or "none"
result.mkcp_mtu = 1350
result.mkcp_tti = 50
result.mkcp_uplinkCapacity = 5
result.mkcp_downlinkCapacity = 20
result.mkcp_readBufferSize = 2
result.mkcp_writeBufferSize = 2
result.mkcp_seed = params.seed
end
if params.type == 'quic' then
result.quic_guise = params.headerType or "none"
result.quic_key = params.key
result.quic_security = params.quicSecurity or "none"
end
if params.type == 'grpc' then
if params.path then result.grpc_serviceName = params.path end
if params.serviceName then result.grpc_serviceName = params.serviceName end
result.grpc_mode = params.mode or "gun"
end
result.tls = "0"
if params.security == "tls" or params.security == "reality" then
result.tls = "1"
result.tls_serverName = (params.sni and params.sni ~= "") and params.sni or params.host
result.alpn = params.alpn
if params.fp and params.fp ~= "" then
result.utls = "1"
result.fingerprint = params.fp
end
if params.ech and params.ech ~= "" then
result.ech = "1"
result.ech_config = params.ech
end
if params.security == "reality" then
result.reality = "1"
result.reality_publicKey = params.pbk or nil
result.reality_shortId = params.sid or nil
result.reality_spiderX = params.spx or nil
result.use_mldsa65Verify = (params.pqv and params.pqv ~= "") and "1" or nil
result.reality_mldsa65Verify = params.pqv or nil
end
end
params.allowinsecure = params.allowinsecure or params.insecure
if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then
result.tls_allowInsecure = params.allowinsecure
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
else
result.error_msg = "请更换Xray或Sing-Box来支持SS更多的传输方式."
end
end
if params["shadow-tls"] then
if result.type ~= "sing-box" and result.type ~= "SS-Rust" then
result.error_msg = ss_type_default .. " 不支持 shadow-tls 插件."
else
-- 解析SS Shadow-TLS 插件参数
local function parseShadowTLSParams(b64str, out)
local ok, data = pcall(jsonParse, base64Decode(b64str))
if not ok or type(data) ~= "table" then return "" end
if type(out) == "table" then
for k, v in pairs(data) do out[k] = v end
end
local t = {}
if data.version then t[#t+1] = "v" .. data.version .. "=1" end
if data.password then t[#t+1] = "passwd=" .. data.password end
for k, v in pairs(data) do
if k ~= "version" and k ~= "password" then
t[#t+1] = k .. "=" .. tostring(v)
end
end
return table.concat(t, ";")
end
if result.type == "SS-Rust" then
result.plugin_enabled = "1"
result.plugin = "shadow-tls"
result.plugin_opts = parseShadowTLSParams(params["shadow-tls"])
elseif result.type == "sing-box" then
local shadowtlsOpt = {}
parseShadowTLSParams(params["shadow-tls"], shadowtlsOpt)
if next(shadowtlsOpt) then
result.shadowtls = "1"
result.shadowtls_version = shadowtlsOpt.version or "1"
result.shadowtls_password = shadowtlsOpt.password
result.shadowtls_serverName = shadowtlsOpt.host
if shadowtlsOpt.fingerprint then
result.shadowtls_utls = "1"
result.shadowtls_fingerprint = shadowtlsOpt.fingerprint or "chrome"
end
end
end
end
end
end
elseif szType == "trojan" then
if trojan_type_default == "trojan-plus" and has_trojan_plus then
result.type = "Trojan-Plus"
elseif trojan_type_default == "sing-box" and has_singbox then
result.type = 'sing-box'
result.protocol = 'trojan'
elseif trojan_type_default == "xray" and has_xray then
result.type = 'Xray'
result.protocol = 'trojan'
else
log("跳过 Trojan 节点,因未适配到 Trojan 核心程序,或未正确设置节点使用类型。")
return nil
end
local alias = ""
if content:find("#") then
local idx_sp = content:find("#")
alias = content:sub(idx_sp + 1, -1)
content = content:sub(0, idx_sp - 1)
end
result.remarks = UrlDecode(alias)
if content:find("@") then
local Info = split(content, "@")
result.password = UrlDecode(Info[1])
local port = "443"
Info[2] = (Info[2] or ""):gsub("/%?", "?")
local query = split(Info[2], "%?")
local host_port = query[1]
local params = {}
for _, v in pairs(split(query[2], '&')) do
local t = split(v, '=')
if #t > 1 then
params[string.lower(t[1])] = UrlDecode(t[2])
end
end
-- [2001:4860:4860::8888]:443
-- 8.8.8.8:443
if host_port:find(":") then
local sp = split(host_port, ":")
port = sp[#sp]
if api.is_ipv6addrport(host_port) then
result.address = api.get_ipv6_only(host_port)
else
result.address = sp[1]
end
else
result.address = host_port
end
local peer, sni = nil, ""
if params.peer then peer = params.peer end
sni = params.sni and params.sni or ""
result.port = port
result.tls = '1'
result.tls_serverName = peer and peer or sni
params.allowinsecure = params.allowinsecure or params.insecure
if params.allowinsecure then
if params.allowinsecure == "1" or params.allowinsecure == "0" then
result.tls_allowInsecure = params.allowinsecure
else
result.tls_allowInsecure = string.lower(params.allowinsecure) == "true" and "1" or "0"
end
--log(result.remarks .. ' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure)
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
if not params.type then params.type = "tcp" end
params.type = string.lower(params.type)
if result.type == "sing-box" and params.type == "raw" then
params.type = "tcp"
elseif result.type == "Xray" and params.type == "tcp" then
params.type = "raw"
end
if params.type == "h2" or params.type == "http" then
params.type = "http"
result.transport = (result.type == "Xray") and "xhttp" or "http"
else
result.transport = params.type
end
if params.type == 'ws' then
result.ws_host = params.host
result.ws_path = params.path
if result.type == "sing-box" and params.path then
local ws_path_dat = split(params.path, "%?")
local ws_path = ws_path_dat[1]
local ws_path_params = {}
for _, v in pairs(split(ws_path_dat[2], '&')) do
local t = split(v, '=')
ws_path_params[t[1]] = t[2]
end
if ws_path_params.ed and tonumber(ws_path_params.ed) then
result.ws_path = ws_path
result.ws_enableEarlyData = "1"
result.ws_maxEarlyData = tonumber(ws_path_params.ed)
result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol"
end
end
end
if params.type == "http" then
if result.type == "sing-box" then
result.transport = "http"
result.http_host = (params.host and params.host ~= "") and { params.host } or nil
result.http_path = params.path
elseif result.type == "Xray" then
result.transport = "xhttp"
result.xhttp_mode = "stream-one"
result.xhttp_host = params.host
result.xhttp_path = params.path
end
end
if params.type == 'raw' or params.type == 'tcp' then
result.tcp_guise = params.headerType or "none"
result.tcp_guise_http_host = (params.host and params.host ~= "") and { params.host } or nil
result.tcp_guise_http_path = (params.path and params.path ~= "") and { params.path } or nil
end
if params.type == 'kcp' or params.type == 'mkcp' then
result.transport = "mkcp"
result.mkcp_guise = params.headerType or "none"
result.mkcp_mtu = 1350
result.mkcp_tti = 50
result.mkcp_uplinkCapacity = 5
result.mkcp_downlinkCapacity = 20
result.mkcp_readBufferSize = 2
result.mkcp_writeBufferSize = 2
result.mkcp_seed = params.seed
end
if params.type == 'quic' then
result.quic_guise = params.headerType or "none"
result.quic_key = params.key
result.quic_security = params.quicSecurity or "none"
end
if params.type == 'grpc' then
if params.path then result.grpc_serviceName = params.path end
if params.serviceName then result.grpc_serviceName = params.serviceName end
result.grpc_mode = params.mode or "gun"
end
if params.type == 'xhttp' then
result.xhttp_host = params.host
result.xhttp_path = params.path
end
if params.type == 'httpupgrade' then
result.httpupgrade_host = params.host
result.httpupgrade_path = params.path
end
result.alpn = params.alpn
if result.type == "sing-box" and (result.transport == "mkcp" or result.transport == "xhttp") then
log("跳过节点:" .. result.remarks ..",因Sing-Box不支持" .. szType .. "协议的" .. result.transport .. "传输方式,需更换Xray。")
return nil
end
end
elseif szType == "ssd" then
result.type = "SS"
result.address = content.server
result.port = content.port
result.password = content.password
result.method = content.encryption
result.plugin = content.plugin
result.plugin_opts = content.plugin_options
result.group = content.airport
result.remarks = content.remarks
elseif szType == "vless" then
if vless_type_default == "sing-box" and has_singbox then
result.type = 'sing-box'
elseif vless_type_default == "xray" and has_xray then
result.type = "Xray"
else
log("跳过 VLESS 节点,因未适配到 VLESS 核心程序,或未正确设置节点使用类型。")
return nil
end
result.protocol = "vless"
local alias = ""
if content:find("#") then
local idx_sp = content:find("#")
alias = content:sub(idx_sp + 1, -1)
content = content:sub(0, idx_sp - 1)
end
result.remarks = UrlDecode(alias)
if content:find("@") then
local Info = split(content, "@")
result.uuid = UrlDecode(Info[1])
local port = "443"
Info[2] = (Info[2] or ""):gsub("/%?", "?")
local query = split(Info[2], "%?")
local host_port = query[1]
local params = {}
for _, v in pairs(split(query[2], '&')) do
local t = split(v, '=')
params[t[1]] = UrlDecode(t[2])
end
-- [2001:4860:4860::8888]:443
-- 8.8.8.8:443
if host_port:find(":") then
local sp = split(host_port, ":")
port = sp[#sp]
if api.is_ipv6addrport(host_port) then
result.address = api.get_ipv6_only(host_port)
else
result.address = sp[1]
end
else
result.address = host_port
end
if not params.type then params.type = "tcp" end
params.type = string.lower(params.type)
if ({ xhttp=true, kcp=true, mkcp=true })[params.type] and result.type ~= "Xray" and has_xray then
result.type = "Xray"
end
if result.type == "sing-box" and params.type == "raw" then
params.type = "tcp"
elseif result.type == "Xray" and params.type == "tcp" then
params.type = "raw"
end
if params.type == "h2" or params.type == "http" then
params.type = "http"
result.transport = (result.type == "Xray") and "xhttp" or "http"
else
result.transport = params.type
end
if params.type == 'ws' then
result.ws_host = params.host
result.ws_path = params.path
if result.type == "sing-box" and params.path then
local ws_path_dat = split(params.path, "%?")
local ws_path = ws_path_dat[1]
local ws_path_params = {}
for _, v in pairs(split(ws_path_dat[2], '&')) do
local t = split(v, '=')
ws_path_params[t[1]] = t[2]
end
if ws_path_params.ed and tonumber(ws_path_params.ed) then
result.ws_path = ws_path
result.ws_enableEarlyData = "1"
result.ws_maxEarlyData = tonumber(ws_path_params.ed)
result.ws_earlyDataHeaderName = "Sec-WebSocket-Protocol"
end
end
end
if params.type == "http" then
if result.type == "sing-box" then
result.transport = "http"
result.http_host = (params.host and params.host ~= "") and { params.host } or nil
result.http_path = params.path
elseif result.type == "Xray" then
result.transport = "xhttp"
result.xhttp_mode = "stream-one"
result.xhttp_host = params.host
result.xhttp_path = params.path
end
end
if params.type == 'raw' or params.type == 'tcp' then
result.tcp_guise = params.headerType or "none"
result.tcp_guise_http_host = (params.host and params.host ~= "") and { params.host } or nil
result.tcp_guise_http_path = (params.path and params.path ~= "") and { params.path } or nil
end
if params.type == 'kcp' or params.type == 'mkcp' then
result.transport = "mkcp"
result.mkcp_guise = params.headerType or "none"
result.mkcp_mtu = 1350
result.mkcp_tti = 50
result.mkcp_uplinkCapacity = 5
result.mkcp_downlinkCapacity = 20
result.mkcp_readBufferSize = 2
result.mkcp_writeBufferSize = 2
result.mkcp_seed = params.seed
end
if params.type == 'quic' then
result.quic_guise = params.headerType or "none"
result.quic_key = params.key
result.quic_security = params.quicSecurity or "none"
end
if params.type == 'grpc' then
if params.path then result.grpc_serviceName = params.path end
if params.serviceName then result.grpc_serviceName = params.serviceName end
result.grpc_mode = params.mode or "gun"
end
if params.type == 'xhttp' then
result.xhttp_host = params.host
result.xhttp_path = params.path
result.xhttp_mode = params.mode or "auto"
result.use_xhttp_extra = (params.extra and params.extra ~= "") and "1" or nil
result.xhttp_extra = (params.extra and params.extra ~= "") and params.extra or nil
local success, Data = pcall(jsonParse, params.extra)
if success and Data then
local address = (Data.extra and Data.extra.downloadSettings and Data.extra.downloadSettings.address)
or (Data.downloadSettings and Data.downloadSettings.address)
result.download_address = (address and address ~= "") and address:gsub("^%[", ""):gsub("%]$", "") or nil
else
result.download_address = nil
end
end
if params.type == 'httpupgrade' then
result.httpupgrade_host = params.host
result.httpupgrade_path = params.path
end
result.encryption = params.encryption or "none"
result.flow = params.flow and params.flow:gsub("-udp443", "") or nil
result.tls = "0"
if params.security == "tls" or params.security == "reality" then
result.tls = "1"
result.tls_serverName = (params.sni and params.sni ~= "") and params.sni or params.host
result.alpn = params.alpn
if params.fp and params.fp ~= "" then
result.utls = "1"
result.fingerprint = params.fp
end
if params.ech and params.ech ~= "" then
result.ech = "1"
result.ech_config = params.ech
end
if params.security == "reality" then
result.reality = "1"
result.reality_publicKey = params.pbk or nil
result.reality_shortId = params.sid or nil
result.reality_spiderX = params.spx or nil
result.use_mldsa65Verify = (params.pqv and params.pqv ~= "") and "1" or nil
result.reality_mldsa65Verify = params.pqv or nil
end
end
result.port = port
params.allowinsecure = params.allowinsecure or params.insecure
if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then
result.tls_allowInsecure = params.allowinsecure
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
if result.type == "sing-box" and (result.transport == "mkcp" or result.transport == "xhttp") then
log("跳过节点:" .. result.remarks ..",因Sing-Box不支持" .. szType .. "协议的" .. result.transport .. "传输方式,需更换Xray。")
return nil
end
end
elseif szType == 'hysteria' then
if has_singbox then
result.type = 'sing-box'
result.protocol = "hysteria"
else
log("跳过 Hysteria 节点,因未安装 Hysteria 核心程序 Sing-box。")
return nil
end
local alias = ""
if content:find("#") then
local idx_sp = content:find("#")
alias = content:sub(idx_sp + 1, -1)
content = content:sub(0, idx_sp - 1)
end
result.remarks = UrlDecode(alias)
local dat = split(content:gsub("/%?", "?"), '%?')
local host_port = dat[1]
local params = {}
for _, v in pairs(split(dat[2], '&')) do
local t = split(v, '=')
if #t > 0 then
params[t[1]] = t[2]
end
end
-- [2001:4860:4860::8888]:443
-- 8.8.8.8:443
if host_port:find(":") then
local sp = split(host_port, ":")
result.port = sp[#sp]
if api.is_ipv6addrport(host_port) then
result.address = api.get_ipv6_only(host_port)
else
result.address = sp[1]
end
else
result.address = host_port
end
result.hysteria_obfs = params.obfsParam
result.hysteria_auth_type = "string"
result.hysteria_auth_password = params.auth
result.tls_serverName = params.peer
params.allowinsecure = params.allowinsecure or params.insecure
if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then
result.tls_allowInsecure = params.allowinsecure
--log(result.remarks ..' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure)
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
result.hysteria_alpn = params.alpn
result.hysteria_up_mbps = params.upmbps
result.hysteria_down_mbps = params.downmbps
result.hysteria_hop = params.mport
elseif szType == 'hysteria2' or szType == 'hy2' then
local alias = ""
if content:find("#") then
local idx_sp = content:find("#")
alias = content:sub(idx_sp + 1, -1)
content = content:sub(0, idx_sp - 1)
end
result.remarks = UrlDecode(alias)
local Info = content
if content:find("@") then
local contents = split(content, "@")
result.hysteria2_auth_password = UrlDecode(contents[1])
Info = (contents[2] or ""):gsub("/%?", "?")
end
local query = split(Info, "%?")
local host_port = query[1]
local params = {}
for _, v in pairs(split(query[2], '&')) do
local t = split(v, '=')
if #t > 1 then
params[string.lower(t[1])] = UrlDecode(t[2])
end
end
-- [2001:4860:4860::8888]:443
-- 8.8.8.8:443
if host_port:find(":") then
local sp = split(host_port, ":")
result.port = sp[#sp]
if api.is_ipv6addrport(host_port) then
result.address = api.get_ipv6_only(host_port)
else
result.address = sp[1]
end
else
result.address = host_port
end
result.tls_serverName = params.sni
params.allowinsecure = params.allowinsecure or params.insecure
if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then
result.tls_allowInsecure = params.allowinsecure
--log(result.remarks ..' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure)
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
result.hysteria2_tls_pinSHA256 = params.pinSHA256
result.hysteria2_hop = params.mport
if hysteria2_type_default == "sing-box" and has_singbox then
result.type = 'sing-box'
result.protocol = "hysteria2"
if params["obfs-password"] or params["obfs_password"] then
result.hysteria2_obfs_type = "salamander"
result.hysteria2_obfs_password = params["obfs-password"] or params["obfs_password"]
end
elseif has_hysteria2 then
result.type = "Hysteria2"
if params["obfs-password"] or params["obfs_password"] then
result.hysteria2_obfs = params["obfs-password"] or params["obfs_password"]
end
else
log("跳过 Hysteria2 节点,因未适配到 Hysteria2 核心程序,或未正确设置节点使用类型。")
return nil
end
elseif szType == 'tuic' then
if has_singbox then
result.type = 'sing-box'
result.protocol = "tuic"
else
log("跳过 Tuic 节点,因未安装 Tuic 核心程序 Sing-box。")
return nil
end
local alias = ""
if content:find("#") then
local idx_sp = content:find("#")
alias = content:sub(idx_sp + 1, -1)
content = content:sub(0, idx_sp - 1)
end
result.remarks = UrlDecode(alias)
local Info = content
if content:find("@") then
local contents = split(content, "@")
if contents[1]:find(":") then
local userinfo = split(contents[1], ":")
result.uuid = UrlDecode(userinfo[1])
result.password = UrlDecode(userinfo[2])
end
Info = (contents[2] or ""):gsub("/%?", "?")
end
local query = split(Info, "%?")
local host_port = query[1]
local params = {}
for _, v in pairs(split(query[2], '&')) do
local t = split(v, '=')
if #t > 1 then
params[string.lower(t[1])] = UrlDecode(t[2])
end
end
if host_port:find(":") then
local sp = split(host_port, ":")
result.port = sp[#sp]
if api.is_ipv6addrport(host_port) then
result.address = api.get_ipv6_only(host_port)
else
result.address = sp[1]
end
else
result.address = host_port
end
result.tls_serverName = params.sni
result.tuic_alpn = params.alpn or "default"
result.tuic_congestion_control = params.congestion_control or "cubic"
result.tuic_udp_relay_mode = params.udp_relay_mode or "native"
params.allowinsecure = params.allowinsecure or params.insecure
if params.allowinsecure then
if params.allowinsecure == "1" or params.allowinsecure == "0" then
result.tls_allowInsecure = params.allowinsecure
else
result.tls_allowInsecure = string.lower(params.allowinsecure) == "true" and "1" or "0"
end
--log(result.remarks .. ' 使用节点AllowInsecure设定: '.. result.tls_allowInsecure)
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
elseif szType == "anytls" then
if has_singbox then
result.type = 'sing-box'
result.protocol = "anytls"
else
log("跳过 AnyTLS 节点,因未安装 AnyTLS 核心程序 Sing-box 1.12。")
return nil
end
local alias = ""
if content:find("#") then
local idx_sp = content:find("#")
alias = content:sub(idx_sp + 1, -1)
content = content:sub(0, idx_sp - 1)
end
result.remarks = UrlDecode(alias)
if content:find("@") then
local Info = split(content, "@")
result.password = UrlDecode(Info[1])
local port = "443"
Info[2] = (Info[2] or ""):gsub("/%?", "?")
local query = split(Info[2], "%?")
local host_port = query[1]
local params = {}
for _, v in pairs(split(query[2], '&')) do
local t = split(v, '=')
params[t[1]] = UrlDecode(t[2])
end
-- [2001:4860:4860::8888]:443
-- 8.8.8.8:443
if host_port:find(":") then
local sp = split(host_port, ":")
port = sp[#sp]
if api.is_ipv6addrport(host_port) then
result.address = api.get_ipv6_only(host_port)
else
result.address = sp[1]
end
else
result.address = host_port
end
result.tls = "0"
if (not params.security or params.security == "") and params.sni and params.sni ~= "" then
params.security = "tls"
end
if params.security == "tls" or params.security == "reality" then
result.tls = "1"
result.tls_serverName = params.sni
result.alpn = params.alpn
if params.fp and params.fp ~= "" then
result.utls = "1"
result.fingerprint = params.fp
end
if params.security == "reality" then
result.reality = "1"
result.reality_publicKey = params.pbk or nil
result.reality_shortId = params.sid or nil
end
end
result.port = port
params.allowinsecure = params.allowinsecure or params.insecure
if params.allowinsecure and (params.allowinsecure == "1" or params.allowinsecure == "0") then
result.tls_allowInsecure = params.allowinsecure
else
result.tls_allowInsecure = allowInsecure_default and "1" or "0"
end
local singbox_version = api.get_app_version("sing-box")
local version_ge_1_12 = api.compare_versions(singbox_version:match("[^v]+"), ">=", "1.12.0")
if not has_singbox or not version_ge_1_12 then
log("跳过节点:" .. result.remarks ..",因" .. szType .. "类型的节点需要 Sing-Box 1.12 以上版本支持。")
return nil
end
end
else
log('暂时不支持' .. szType .. "类型的节点订阅,跳过此节点。")
return nil
end
if not result.remarks or result.remarks == "" then
if result.address and result.port then
result.remarks = result.address .. ':' .. result.port
else
result.remarks = "NULL"
end
end
return result
end
local function curl(url, file, ua, mode)
local curl_args = {
"-skL", "-w %{http_code}", "--retry 3", "--connect-timeout 3"
}
if ua and ua ~= "" and ua ~= "curl" then
curl_args[#curl_args + 1] = '--user-agent "' .. ua .. '"'
end
local return_code, result
if mode == "direct" then
return_code, result = api.curl_direct(url, file, curl_args)
elseif mode == "proxy" then
return_code, result = api.curl_proxy(url, file, curl_args)
else
return_code, result = api.curl_auto(url, file, curl_args)
end
return tonumber(result)
end
local function truncate_nodes(add_from)
for _, config in pairs(CONFIG) do
if config.currentNodes and #config.currentNodes > 0 then
local newNodes = {}
local removeNodesSet = {}
for k, v in pairs(config.currentNodes) do
if v.currentNode and v.currentNode.add_mode == "2" then
if (not add_from) or (add_from and add_from == v.currentNode.add_from) then
removeNodesSet[v.currentNode[".name"]] = true
end
end
end
for _, value in ipairs(config.currentNodes) do
if not removeNodesSet[value.currentNode[".name"]] then
newNodes[#newNodes + 1] = value.currentNode[".name"]
end
end
if config.set then
config.set(config, newNodes)
end
else
if config.currentNode and config.currentNode.add_mode == "2" then
if (not add_from) or (add_from and add_from == config.currentNode.add_from) then
if config.delete then
config.delete(config)
elseif config.set then
config.set(config, "")
end
end
end
end
end
uci:foreach(appname, "nodes", function(node)
if node.add_mode == "2" then
if (not add_from) or (add_from and add_from == node.add_from) then
uci:delete(appname, node['.name'])
end
end
end)
uci:foreach(appname, "subscribe_list", function(o)
if (not add_from) or add_from == o.remark then
uci:delete(appname, o['.name'], "md5")
end
end)
api.uci_save(uci, appname, true)
end
local function select_node(nodes, config, parentConfig)
if config.currentNode then
local server
-- 特别优先级 cfgid
if config.currentNode[".name"] then
for index, node in pairs(nodes) do
if node[".name"] == config.currentNode[".name"] then
if config.log == nil or config.log == true then
log('更新【' .. config.remarks .. '】匹配节点:' .. node.remarks)
end
server = node[".name"]
break
end
end
end
-- 第一优先级 类型 + 备注 + IP + 端口
if not server then
for index, node in pairs(nodes) do
if config.currentNode.type and config.currentNode.remarks and config.currentNode.address and config.currentNode.port then
if node.type and node.remarks and node.address and node.port then
if node.type == config.currentNode.type and node.remarks == config.currentNode.remarks and (node.address .. ':' .. node.port == config.currentNode.address .. ':' .. config.currentNode.port) then
if config.log == nil or config.log == true then
log('更新【' .. config.remarks .. '】第一匹配节点:' .. node.remarks)
end
server = node[".name"]
break
end
end
end
end
end
-- 第二优先级 类型 + IP + 端口
if not server then
for index, node in pairs(nodes) do
if config.currentNode.type and config.currentNode.address and config.currentNode.port then
if node.type and node.address and node.port then
if node.type == config.currentNode.type and (node.address .. ':' .. node.port == config.currentNode.address .. ':' .. config.currentNode.port) then
if config.log == nil or config.log == true then
log('更新【' .. config.remarks .. '】第二匹配节点:' .. node.remarks)
end
server = node[".name"]
break
end
end
end
end
end
-- 第三优先级 IP + 端口
if not server then
for index, node in pairs(nodes) do
if config.currentNode.address and config.currentNode.port then
if node.address and node.port then
if node.address .. ':' .. node.port == config.currentNode.address .. ':' .. config.currentNode.port then
if config.log == nil or config.log == true then
log('更新【' .. config.remarks .. '】第三匹配节点:' .. node.remarks)
end
server = node[".name"]
break
end
end
end
end
end
-- 第四优先级 IP
if not server then
for index, node in pairs(nodes) do
if config.currentNode.address then
if node.address then
if node.address == config.currentNode.address then
if config.log == nil or config.log == true then
log('更新【' .. config.remarks .. '】第四匹配节点:' .. node.remarks)
end
server = node[".name"]
break
end
end
end
end
end
-- 第五优先级备注
if not server then
for index, node in pairs(nodes) do
if config.currentNode.remarks then
if node.remarks then
if node.remarks == config.currentNode.remarks then
if config.log == nil or config.log == true then
log('更新【' .. config.remarks .. '】第五匹配节点:' .. node.remarks)
end
server = node[".name"]
break
end
end
end
end
end
if not parentConfig then
-- 还不行 随便找一个
if not server then
if #nodes_table > 0 then
if config.log == nil or config.log == true then
log('【' .. config.remarks .. '】' .. '无法找到最匹配的节点,当前已更换为:' .. nodes_table[1].remarks)
end
server = nodes_table[1][".name"]
end
end
end
if server then
if parentConfig then
config.set(parentConfig, server)
else
config.set(config, server)
end
end
else
if not parentConfig then
config.set(config, "")
end
end
end
local function update_node(manual)
if next(nodeResult) == nil then
log("没有可用的节点信息更新。")
return
end
local group = {}
for _, v in ipairs(nodeResult) do
group[v["remark"]] = true
end
if manual == 0 and next(group) then
uci:foreach(appname, "nodes", function(node)
-- 如果未发现新节点或手动导入的节点就不要删除了...
if node.add_mode == "2" and (node.add_from and group[node.add_from] == true) then
uci:delete(appname, node['.name'])
end
end)
end
for _, v in ipairs(nodeResult) do
local remark = v["remark"]
local list = v["list"]
for _, vv in ipairs(list) do
local cfgid = uci:section(appname, "nodes", api.gen_short_uuid())
for kkk, vvv in pairs(vv) do
if type(vvv) == "table" and next(vvv) ~= nil then
uci:set_list(appname, cfgid, kkk, vvv)
else
uci:set(appname, cfgid, kkk, vvv)
-- sing-box 域名解析策略
if kkk == "type" and vvv == "sing-box" then
uci:set(appname, cfgid, "domain_strategy", domain_strategy_node)
end
-- 订阅组链式代理
if chain_node_type ~= "" and kkk == "type" and vvv == chain_node_type then
if preproxy_node_group ~="" then
uci:set(appname, cfgid, "chain_proxy", "1")
uci:set(appname, cfgid, "preproxy_node", preproxy_node_group)
elseif to_node_group ~= "" then
uci:set(appname, cfgid, "chain_proxy", "2")
uci:set(appname, cfgid, "to_node", to_node_group)
end
end
end
end
end
end
-- 更新机场信息
for cfgid, info in pairs(subscribe_info) do
for key, value in pairs(info) do
if value ~= "" then
uci:set(appname, cfgid, key, value)
else
uci:delete(appname, cfgid, key)
end
end
end
api.uci_save(uci, appname, true)
if next(CONFIG) then
local nodes = {}
uci:foreach(appname, "nodes", function(node)
nodes[#nodes + 1] = node
end)
for _, config in pairs(CONFIG) do
if config.currentNodes and #config.currentNodes > 0 then
if config.remarks and config.currentNodes[1].log ~= false then
log('----【' .. config.remarks .. '】----')
end
for kk, vv in pairs(config.currentNodes) do
select_node(nodes, vv, config)
end
config.set(config)
else
select_node(nodes, config)
end
end
api.uci_save(uci, appname, true)
end
if arg[3] == "cron" then
if not fs.access("/var/lock/" .. appname .. ".lock") then
luci.sys.call("touch /tmp/lock/" .. appname .. "_cron.lock")
end
end
luci.sys.call("/etc/init.d/" .. appname .. " restart > /dev/null 2>&1 &")
end
local function parse_link(raw, add_mode, add_from, cfgid)
if raw and #raw > 0 then
local nodes, szType
local node_list = {}
-- SSD 似乎是这种格式 ssd:// 开头的
if raw:find('ssd://') then
szType = 'ssd'
local nEnd = select(2, raw:find('ssd://'))
nodes = base64Decode(raw:sub(nEnd + 1, #raw))
nodes = jsonParse(nodes)
local extra = {
airport = nodes.airport,
port = nodes.port,
encryption = nodes.encryption,
password = nodes.password
}
local servers = {}
-- SS里面包着 干脆直接这样
for _, server in ipairs(nodes.servers) do
tinsert(servers, setmetatable(server, { __index = extra }))
end
nodes = servers
else
-- ssd 外的格式
if add_mode == "1" then
nodes = split(raw, "\n")
else
nodes = split(base64Decode(raw):gsub("\r\n", "\n"), "\n")
end
end
for _, v in ipairs(nodes) do
if v and not string.match(v, "^%s*$") then
xpcall(function ()
local result
if szType == 'ssd' then
result = processData(szType, v, add_mode, add_from)
elseif not szType then
local node = api.trim(v)
local dat = split(node, "://")
if dat and dat[1] and dat[2] then
if dat[1] == 'vmess' or dat[1] == 'ssr' then
local link = api.trim(dat[2]:gsub("#.*$", ""))
result = processData(dat[1], base64Decode(link), add_mode, add_from)
else
local link = dat[2]:gsub("&", "&"):gsub("%s*#%s*", "#") -- 一些奇葩的链接用"&"当做"&","#"前后带空格
result = processData(dat[1], link, add_mode, add_from)
end
end
else
log('跳过未知类型: ' .. szType)
end
-- log(result)
if result then
if result.error_msg then
log('丢弃节点: ' .. result.remarks .. ", 原因:" .. result.error_msg)
elseif not result.type then
log('丢弃节点: ' .. result.remarks .. ", 找不到可使用二进制.")
elseif (add_mode == "2" and is_filter_keyword(result.remarks)) or not result.address or result.remarks == "NULL" or result.address == "127.0.0.1" or
(not datatypes.hostname(result.address) and not (api.is_ip(result.address))) then
log('丢弃过滤节点: ' .. result.type .. ' 节点, ' .. result.remarks)
else
tinsert(node_list, result)
end
if add_mode == "2" then
get_subscribe_info(cfgid, result.remarks)
end
end
end, function (err)
--log(err)
log(v, "解析错误,跳过此节点。")
end
)
end
end
if #node_list > 0 then
nodeResult[#nodeResult + 1] = {
remark = add_from,
list = node_list
}
end
log('成功解析【' .. add_from .. '】节点数量: ' .. #node_list)
else
if add_mode == "2" then
log('获取到的【' .. add_from .. '】订阅内容为空,可能是订阅地址无效,或是网络问题,请诊断!')
end
end
end
local execute = function()
do
local subscribe_list = {}
local fail_list = {}
if arg[2] ~= "all" then
string.gsub(arg[2], '[^' .. "," .. ']+', function(w)
subscribe_list[#subscribe_list + 1] = uci:get_all(appname, w) or {}
end)
else
uci:foreach(appname, "subscribe_list", function(o)
subscribe_list[#subscribe_list + 1] = o
end)
end
local manual_sub = arg[3] == "manual"
for index, value in ipairs(subscribe_list) do
local cfgid = value[".name"]
local remark = value.remark
local url = value.url
if value.allowInsecure and value.allowInsecure == "1" then
allowInsecure_default = true
end
local filter_keyword_mode = value.filter_keyword_mode or "5"
if filter_keyword_mode == "0" then
filter_keyword_mode_default = "0"
elseif filter_keyword_mode == "1" then
filter_keyword_mode_default = "1"
filter_keyword_discard_list_default = value.filter_discard_list or {}
elseif filter_keyword_mode == "2" then
filter_keyword_mode_default = "2"
filter_keyword_keep_list_default = value.filter_keep_list or {}
elseif filter_keyword_mode == "3" then
filter_keyword_mode_default = "3"
filter_keyword_keep_list_default = value.filter_keep_list or {}
filter_keyword_discard_list_default = value.filter_discard_list or {}
elseif filter_keyword_mode == "4" then
filter_keyword_mode_default = "4"
filter_keyword_keep_list_default = value.filter_keep_list or {}
filter_keyword_discard_list_default = value.filter_discard_list or {}
end
local ss_type = value.ss_type or "global"
if ss_type ~= "global" then
ss_type_default = ss_type
end
local trojan_type = value.trojan_type or "global"
if trojan_type ~= "global" then
trojan_type_default = trojan_type
end
local vmess_type = value.vmess_type or "global"
if vmess_type ~= "global" then
vmess_type_default = vmess_type
end
local vless_type = value.vless_type or "global"
if vless_type ~= "global" then
vless_type_default = vless_type
end
local hysteria2_type = value.hysteria2_type or "global"
if hysteria2_type ~= "global" then
hysteria2_type_default = hysteria2_type
end
local domain_strategy = value.domain_strategy or "global"
if domain_strategy ~= "global" then
domain_strategy_node = domain_strategy
else
domain_strategy_node = domain_strategy_default
end
-- 订阅组链式代理
local function valid_chain_node(node)
if not node then return "" end
local cp = uci:get(appname, node, "chain_proxy") or ""
local am = uci:get(appname, node, "add_mode") or "0"
chain_node_type = (cp == "" and am ~= "2") and (uci:get(appname, node, "type") or "") or ""
if chain_node_type ~= "Xray" and chain_node_type ~= "sing-box" then
chain_node_type = ""
return ""
end
return node
end
preproxy_node_group = (value.chain_proxy == "1") and valid_chain_node(value.preproxy_node) or ""
to_node_group = (value.chain_proxy == "2") and valid_chain_node(value.to_node) or ""
local ua = value.user_agent
local access_mode = value.access_mode
local result = (not access_mode) and "自动" or (access_mode == "direct" and "直连访问" or (access_mode == "proxy" and "通过代理" or "自动"))
log('正在订阅:【' .. remark .. '】' .. url .. ' [' .. result .. ']')
local tmp_file = "/tmp/" .. cfgid
value.http_code = curl(url, tmp_file, ua, access_mode)
if value.http_code ~= 200 then
fail_list[#fail_list + 1] = value
else
if luci.sys.call("[ -f " .. tmp_file .. " ] && sed -i -e '/^[ \t]*$/d' -e '/^[ \t]*\r$/d' " .. tmp_file) == 0 then
local f = io.open(tmp_file, "r")
local stdout = f:read("*all")
f:close()
local raw_data = api.trim(stdout)
local old_md5 = value.md5 or ""
local new_md5 = luci.sys.exec("md5sum " .. tmp_file .. " 2>/dev/null | awk '{print $1}'"):gsub("\n", "")
if not manual_sub and old_md5 == new_md5 then
log('订阅:【' .. remark .. '】没有变化,无需更新。')
else
parse_link(raw_data, "2", remark, cfgid)
uci:set(appname, cfgid, "md5", new_md5)
end
else
fail_list[#fail_list + 1] = value
end
end
luci.sys.call("rm -f " .. tmp_file)
allowInsecure_default = nil
filter_keyword_mode_default = uci:get(appname, "@global_subscribe[0]", "filter_keyword_mode") or "0"
filter_keyword_discard_list_default = uci:get(appname, "@global_subscribe[0]", "filter_discard_list") or {}
filter_keyword_keep_list_default = uci:get(appname, "@global_subscribe[0]", "filter_keep_list") or {}
ss_type_default = uci:get(appname, "@global_subscribe[0]", "ss_type") or "shadowsocks-libev"
trojan_type_default = uci:get(appname, "@global_subscribe[0]", "trojan_type") or "trojan-plus"
vmess_type_default = uci:get(appname, "@global_subscribe[0]", "vmess_type") or "xray"
vless_type_default = uci:get(appname, "@global_subscribe[0]", "vless_type") or "xray"
hysteria2_type_default = uci:get(appname, "@global_subscribe[0]", "hysteria2_type") or "hysteria2"
end
if #fail_list > 0 then
for index, value in ipairs(fail_list) do
log(string.format('【%s】订阅失败,可能是订阅地址无效,或是网络问题,请诊断![%s]', value.remark, tostring(value.http_code)))
end
end
update_node(0)
end
end
if arg[1] then
if arg[1] == "start" then
log('开始订阅...')
xpcall(execute, function(e)
log(e)
if type(debug) == "table" and type(debug.traceback) == "function" then
log(debug.traceback())
end
log('发生错误, 正在恢复服务')
end)
log('订阅完毕...\n')
elseif arg[1] == "add" then
local f = assert(io.open("/tmp/links.conf", 'r'))
local raw = f:read('*all')
f:close()
parse_link(raw, "1", "导入")
update_node(1)
luci.sys.call("rm -f /tmp/links.conf")
elseif arg[1] == "truncate" then
truncate_nodes(arg[2])
end
end
|
294coder/Efficient-MIF | 2,476 | configs/LEMamba_config.yaml | optimizer:
name: "adamw"
lr: !!float 1e-3
weight_decay: !!float 1e-6
lr_scheduler:
name: 'cos_anneal_restart_reduce'
T_0: 50
T_mult: 2
lr_mult: 0.9
eta_min: !!float 8e-5
max_norm: 0.03
ema_decay: 0.995
path:
wv3_train_path: <your_path>/train_data/
wv3_val_path: <your_path>/val_data/
gf2_train_path: <your_path>/train_data/
gf2_val_path: <your_path>/val_data/
cave_x8_train_path: <your_path>/train_data/
cave_x8_val_path: <your_path>/val_data/
harvard_x8_train_path: <your_path>/train_data/
harvard_x8_val_path: <your_path>/val_data/
network_configs:
LEMamba:
img_channel: 4
condition_channel: 1
out_channel: 4
# v2
# width: 32
naf_enc_blk_nums: []
naf_dec_blk_nums: []
naf_chan_upscale: []
middle_blk_nums: 1
# variant tiny (wv3)
width: 32
ssm_enc_blk_nums: [2, 1, 1]
ssm_dec_blk_nums: [2, 1, 1]
ssm_chan_upscale: [1, 1, 1]
ssm_ratios: [2, 2, 1]
window_sizes: [8,8,null]
ssm_enc_d_states: [[16, 32], [16, 32], [null, 32]]
ssm_dec_d_states: [[null, 32], [16, 32], [16, 32]]
ssm_enc_convs: [[7, 11], [7, 11], [null, 11]]
ssm_dec_convs: [[null, 11], [7, 11], [7, 11]]
drop_path_rate: 0.1
# small
# ssm_enc_blk_nums: [2,2,2]
# ssm_dec_blk_nums: [2,2,2]
# ssm_chan_upscale: [1,2,2]
# ssm_ratios: [3,2,2]
# window_sizes: [8,8,null]
# ssm_enc_d_states: [[32, 32], [32, 32], [null, 32]]
# ssm_dec_d_states: [[null, 32], [32, 32], [32, 32]]
# ssm_enc_convs: [[7, 11], [7, 11], [null, 11]]
# ssm_dec_convs: [[null, 11], [7, 11], [null, 11]]
# drop_path_rate: 0.2
# base
# ssm_enc_blk_nums: [4, 3, 2]
# ssm_dec_blk_nums: [4, 3, 2]
# ssm_chan_upscale: [1, 1, 1]
# ssm_ratios: [2,2,1]
# window_sizes: [8,8,null]
# ssm_enc_d_states: [[16, 32], [16, 32], [null, 32]]
# ssm_dec_d_states: [[null, 32], [16, 32], [16, 32]]
# ssm_enc_convs: [[7, 11], [7, 11], [null, 11]]
# ssm_dec_convs: [[null, 11], [7, 11], [7, 11]]
# drop_path_rate: 0.2
if_abs_pos: no
if_rope: no
patch_merge: yes
upscale: 4
LEMamba_only_NAF:
img_channel: 8
condition_channel: 1
out_channel: 8
width: 32
naf_enc_blk_nums: [2, 2, 2]
naf_dec_blk_nums: [2, 2, 2]
naf_chan_upscale: [2, 2, 2]
middle_blk_nums: 2
if_abs_pos: no
if_rope: no
patch_merge: yes
upscale: 4
logger_config:
base_path: ./log_file/
name: LEMamba
file_mode: w
|
294coder/Efficient-MIF | 1,569 | configs/dct_config.yaml | optimizer:
name: "adamw"
lr: !!float 1e-3
weight_decay: !!float 1e-6
lr_scheduler:
name: "multi_step"
milestones: [200, 1600]
gamma: 0.1
# name: 'identity'
# name: "plateau"
# mode: "min"
# threshold: !!float 5e-6 # l1
# min_lr: !!float 1e-6
# patience: 30
# factor: !!float 0.2
# cooldown: 10
max_norm: 0.03
ema_decay: 0.999
path:
wv3_train_path: "/volsparse1/dataset/PanCollection/training_data/train_wv3_9714.h5"
wv3_val_path: "/volsparse1/dataset/PanCollection/test_data/test_wv3_multiExm1.h5"
gf2_train_path: "/volsparse1/dataset/PanCollection/training_data/train_gf2_19809.h5"
gf2_val_path: "/volsparse1/dataset/PanCollection/test_data/test_gf2_multiExm1.h5"
cave_x4_train_path: "/volsparse1/dataset/HISR/cave_x4/train_cave(with_up)x4.h5"
cave_x4_val_path: "/volsparse1/dataset/HISR/cave_x4/test_cave(with_up)x4.h5"
cave_x8_train_path: "/volsparse1/dataset/HISR/cave_x8/train_cave(with_up)x8_rgb_16.h5"
cave_x8_val_path: "/volsparse1/dataset/HISR/cave_x8/test_cave(with_up)x8_rgb.h5"
harvard_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5"
harvard_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/test_harvard(with_up)x4_rgb.h5"
harvard_x8_train_path: "/volsparse1/dataset/HISR/harvard_x8/train_harvard(with_up)x8_rgb.h5"
harvard_x8_val_path: "/volsparse1/dataset/HISR/harvard_x8/test_harvard(with_up)x8_rgb.h5"
network_configs:
n_colors: 31
upscale_factor: 8
logger_config:
base_path: ./log_file/
name: reciprocal_transformer
file_mode: w
|
2977094657/DidaAPI | 4,417 | routers/habits.py | """习惯管理相关API路由"""
from fastapi import APIRouter
from fastapi.responses import Response
# 不再需要响应模型导入
from services import habit_service, dida_service
from utils import app_logger
router = APIRouter(prefix="/habits", tags=["习惯管理"])
@router.get("/all",
summary="获取所有习惯",
description="获取当前用户的所有习惯列表")
async def get_all_habits():
"""
获取所有习惯
返回当前用户的所有习惯列表,包括:
- 习惯ID、名称、图标、颜色
- 习惯状态、激励语句、总打卡次数
- 创建时间、修改时间、类型、目标值等信息
**注意**: 需要先完成微信登录获取认证会话
"""
try:
app_logger.info("请求获取所有习惯")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"}
# 获取认证信息
current_session = dida_service.current_session
auth_token = current_session['auth_token']
csrf_token = current_session['csrf_token']
# 调用习惯服务
result = await habit_service.get_habits(auth_token, csrf_token)
if not result:
return {"error": "service_error", "message": "获取习惯列表失败,请稍后重试"}
# 记录日志
if 'error' in result:
app_logger.info(f"习惯获取失败: {result.get('error')}")
else:
habit_count = len(result) if isinstance(result, list) else 0
app_logger.info(f"习惯获取完成,习惯数: {habit_count}")
# 直接返回原始响应
return result
except Exception as e:
app_logger.error(f"获取习惯时发生未知错误: {e}")
return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
@router.get("/statistics/week/current",
summary="获取本周习惯打卡统计",
description="获取本周的习惯打卡统计信息")
async def get_week_current_statistics():
"""
获取本周习惯打卡统计
返回本周的习惯打卡统计,包括:
- 每日打卡情况
- 习惯完成率
- 连续打卡天数等信息
**注意**: 需要先完成微信登录获取认证会话
"""
try:
app_logger.info("请求获取本周习惯打卡统计")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"}
# 获取认证信息
current_session = dida_service.current_session
auth_token = current_session['auth_token']
csrf_token = current_session['csrf_token']
# 调用习惯服务
result = await habit_service.get_week_current_statistics(auth_token, csrf_token)
if not result:
return {"error": "service_error", "message": "获取本周习惯打卡统计失败,请稍后重试"}
# 记录日志
if 'error' in result:
app_logger.info(f"本周习惯打卡统计获取失败: {result.get('error')}")
else:
app_logger.info(f"本周习惯打卡统计获取完成")
# 直接返回原始响应
return result
except Exception as e:
app_logger.error(f"获取本周习惯打卡统计时发生未知错误: {e}")
return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
@router.get("/export",
summary="导出习惯数据",
description="导出习惯数据为Excel文件")
async def export_habits():
"""
导出习惯数据(Excel格式)
导出当前用户的所有习惯数据为Excel文件,包括:
- 习惯基本信息
- 打卡记录
- 统计数据等
**注意**: 需要先完成微信登录获取认证会话
"""
try:
app_logger.info("请求导出习惯数据")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"}
# 获取认证信息
current_session = dida_service.current_session
auth_token = current_session['auth_token']
csrf_token = current_session['csrf_token']
# 调用习惯服务
result = await habit_service.export_habits(auth_token, csrf_token)
if not result:
return {"error": "service_error", "message": "导出习惯数据失败,请稍后重试"}
# 检查是否有错误
if 'error' in result:
app_logger.info(f"习惯数据导出失败: {result.get('error')}")
return result
# 成功获取文件内容
app_logger.info(f"习惯数据导出完成,文件名: {result.get('filename')}")
# 返回文件响应
return Response(
content=result['content'],
media_type=result['content_type'],
headers={
"Content-Disposition": f"attachment; filename={result['filename']}"
}
)
except Exception as e:
app_logger.error(f"导出习惯数据时发生未知错误: {e}")
return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
|
2977094657/DidaAPI | 1,528 | routers/projects.py | """清单管理相关API路由"""
from fastapi import APIRouter
from services import project_service, dida_service
from utils import app_logger
router = APIRouter(prefix="/projects", tags=["清单管理"])
@router.get("/all",
summary="获取所有项目/清单",
description="获取当前用户的所有项目/清单列表")
async def get_all_projects():
"""
获取所有项目/清单
返回当前用户的所有项目/清单列表,包括:
- 项目ID、名称、颜色
- 项目权限、类型、用户数量
- 创建时间、修改时间等信息
**注意**: 需要先完成微信登录获取认证会话
"""
try:
app_logger.info("请求获取所有项目")
# 检查认证状态
session_status = dida_service.get_session_status()
if not session_status["has_session"]:
return {"error": "no_auth_session", "message": "未设置认证会话,请先完成微信登录"}
# 获取认证信息
current_session = dida_service.current_session
auth_token = current_session['auth_token']
csrf_token = current_session['csrf_token']
# 调用项目服务
result = await project_service.get_projects(auth_token, csrf_token)
if not result:
return {"error": "service_error", "message": "获取项目列表失败,请稍后重试"}
# 记录日志
if 'error' in result:
app_logger.info(f"项目获取失败: {result.get('error')}")
else:
project_count = len(result) if isinstance(result, list) else 0
app_logger.info(f"项目获取完成,项目数: {project_count}")
# 直接返回原始响应
return result
except Exception as e:
app_logger.error(f"获取项目时发生未知错误: {e}")
return {"error": "server_error", "message": f"服务器内部错误: {str(e)}"}
|
294coder/Efficient-MIF | 2,223 | configs/MIMO_SST_config.yaml | optimizer:
name: "adamw"
lr: !!float 1e-4
weight_decay: !!float 1e-6
lr_scheduler:
# name: 'cos_anneal'
# name: 'cos_anneal_restart'
# T_0: 50
# T_mult: 2
# eta_min: !!float 1e-9
# T_max: 2000
# eta_min: !!float 1e-8
# name: "plateau"
# mode: "min"
# threshold: !!float 1e-5 # l1
# min_lr: !!float 1e-6
# patience: 20
# factor: !!float 0.1
# cooldown: 5
# threshold_mode: "rel"
name: 'identity'
# name: "multi_step"
# milestones: [1000] #[100, 800]
# gamma: 0.2
max_norm: 0.03
ema_decay: 0.999
path:
# pansharpening
wv3_train_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5"
wv3_val_path: "/Data2/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5"
gf2_train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5"
gf2_val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5"
qb_train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5"
qb_val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"
# hyperspectral multispectral image fusion
cave_x4_train_path: "/volsparse1/dataset/HISR/cave_x8/train_cave(with_up)x8_rgb_16.h5"
# val_path: "/home/ShangqiDeng/data/HSI/cave_x4/test_cavepatches64-2.h5"
cave_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5"
cave_x8_train_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/train_cave(with_up)x8_rgb_16.h5"
# val_path: "/home/ZiHanCao/datasets/HISI/new_cave/x8/validation_cave(with_up)x8_rgb.h5"
cave_x8_val_path: "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/test_cave(with_up)x8_rgb.h5"
harvard_x4_train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5"
harvard_x4_val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/test_harvard(with_up)x4_rgb.h5"
harvard_x8_train_path: "/volsparse1/dataset/HISR/harvard_x8/train_harvard(with_up)x8_rgb.h5"
harvard_x8_val_path: "/volsparse1/dataset/HISR/harvard_x8/test_harvard(with_up)x8_rgb.h5"
network_configs:
in_chan: 8
logger_config:
base_path: ./log_file/
name: MIM_SST
file_mode: w
|
294coder/Efficient-MIF | 3,581 | configs/pmacnet_config.yaml | optimizer:
name: "adamw"
lr: !!float 3e-4
weight_decay: !!float 1e-6
lr_scheduler:
# name: 'multi_step'
# milestones: [400, 800]
# gamma: 0.1
name: "identity"
max_norm: null
path:
# train_path: '/Data2/DataSet/HISR/CAVE/train_cave(with_up)x4.h5'
# val_path: '/Data2/DataSet/HISR/CAVE/validation_cave(with_up)x4.h5'
# train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/training_wv3/train_wv3.h5'
# val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5'
# train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5'
# val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/training_gf2/valid_gf2.h5'
# train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5'
# val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/training_qb/valid_qb.h5'
# train_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5"
# val_path: "/Data2/ZiHanCao/datasets/pansharpening/gf/training_gf2/valid_gf2.h5"
train_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"
val_path: "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"
# train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5'
# val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5'
# train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5'
# val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5'
# train_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5'
# val_path: '/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5'
# train_path: '/home/ZiHanCao/datasets/pansharpening/gf/training_gf2/train_gf2.h5'
# val_path: '/home/ZiHanCao/datasets/pansharpening/gf/training_gf2/valid_gf2.h5'
# train_path: '/home/ZiHanCao/datasets/pansharpening/qb/training_qb/train_qb.h5'
# val_path: '/home/ZiHanCao/datasets/pansharpening/qb/training_qb/valid_qb.h5'
# train_path: "/media/office-401-remote/Elements SE/cao/ZiHanCao/datasets/HISI/new_cave/train_cave(with_up)x4.h5"
# val_path: "/media/office-401-remote/Elements SE/cao/cavex4/test_cavepatches64-2.h5"
# train_path: "/home/ShangqiDeng/data/HSI/cave_x8/train_cave(with_up)x8_rgb.h5"
# val_path: "/home/ZiHanCao/datasets/HISI/new_cave/x8/validation_cave(with_up)x8_rgb.h5"
# train_path: "/home/ShangqiDeng/data/HSI/cave_x8/train_cave(with_up)x8_rgb_16.h5"
# val_path: "/home/ShangqiDeng/data/HSI/cave_x8/test_cavepatches128-2.h5"
# train_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/train_harvard(with_up)x8_rgb.h5"
# val_path: "/Data2/ZiHanCao/datasets/HISI/new_harvard/x8/test_clip_128.h5"
# train_path: '/home/ZiHanCao/datasets/HISI/new_harvard/train_harvard(with_up)x4_rgb.h5'
# val_path: '/home/ZiHanCao/datasets/HISI/new_harvard/validation_harvard(with_up)x4_rgb.h5'
network_configs:
ms_inp_ch: 4
num_layers: 4
latent_dim: 64
logger_config:
base_path: ./log_file/
name: pmacnet
file_mode: w
|
2977094657/DidaAPI | 7,590 | routers/auth.py | """认证相关API路由"""
from fastapi import APIRouter, HTTPException, Query
from fastapi.responses import FileResponse
from models import WeChatQRResponse, WeChatValidateResponse, ApiResponse, PasswordLoginRequest
from services import wechat_service
from utils import app_logger
import os
router = APIRouter(prefix="/auth", tags=["认证"])
@router.get("/wechat/login",
summary="微信扫码登录页面",
description="返回完整的微信扫码登录HTML页面")
async def wechat_login_page():
"""
微信扫码登录页面
返回一个完整的HTML页面,集成了:
- 自动获取微信二维码
- 实时显示二维码
- 自动轮询登录状态
- 登录成功处理
- 错误处理和重试
用户只需访问此页面即可完成完整的微信扫码登录流程
"""
try:
static_file = os.path.join("static", "wechat_login.html")
if not os.path.exists(static_file):
raise HTTPException(
status_code=404,
detail="微信登录页面文件不存在"
)
return FileResponse(
static_file,
media_type="text/html",
headers={"Cache-Control": "no-cache"}
)
except Exception as e:
app_logger.error(f"返回微信登录页面时发生错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/wechat/qrcode",
response_model=WeChatQRResponse,
summary="获取微信登录二维码",
description="获取微信登录二维码,用户扫码后可进行登录验证")
async def get_wechat_qrcode(
state: str = Query(default="Lw==", description="状态参数,用于防止CSRF攻击")
) -> WeChatQRResponse:
"""
获取微信登录二维码
- **state**: 状态参数,默认为 "Lw==",用于防止CSRF攻击
返回包含二维码图片URL和密钥的响应
"""
try:
app_logger.info(f"请求获取微信二维码,state: {state}")
qr_response = await wechat_service.get_qr_code(state)
if not qr_response:
raise HTTPException(
status_code=500,
detail="获取微信二维码失败,请稍后重试"
)
app_logger.info(f"成功返回微信二维码: {qr_response.qr_code_key}")
return qr_response
except HTTPException:
raise
except Exception as e:
app_logger.error(f"获取微信二维码时发生未知错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/wechat/validate",
response_model=WeChatValidateResponse,
summary="验证微信登录",
description="使用扫码后获得的验证码进行微信登录验证")
async def validate_wechat_login(
code: str = Query(..., description="扫码后获得的验证码"),
state: str = Query(default="Lw==", description="状态参数,需与获取二维码时的参数一致")
) -> WeChatValidateResponse:
"""
验证微信登录
- **code**: 扫码后获得的验证码(必需)
- **state**: 状态参数,需与获取二维码时的参数一致
返回登录验证结果,包含用户信息和认证令牌
"""
try:
app_logger.info(f"请求验证微信登录,code: {code}, state: {state}")
if not code:
raise HTTPException(
status_code=400,
detail="验证码不能为空"
)
validate_response = await wechat_service.validate_wechat_login(code, state)
if not validate_response:
raise HTTPException(
status_code=500,
detail="验证微信登录失败,请稍后重试"
)
app_logger.info(f"微信登录验证完成,成功: {validate_response.success}")
return validate_response
except HTTPException:
raise
except Exception as e:
app_logger.error(f"验证微信登录时发生未知错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/wechat/poll",
summary="轮询微信登录状态",
description="轮询检查二维码是否已被扫码登录")
async def poll_wechat_login(
qr_code_key: str = Query(..., description="二维码密钥(16位字符)"),
max_attempts: int = Query(default=60, description="最大轮询次数,默认60次(约5分钟)")
) -> WeChatValidateResponse:
"""
轮询微信登录状态
- **qr_code_key**: 二维码密钥(16位字符)
- **max_attempts**: 最大轮询次数,默认60次(约5分钟)
这个接口会持续轮询微信服务器,检查二维码是否已被扫码登录
"""
try:
app_logger.info(f"开始轮询微信登录状态,qr_code_key: {qr_code_key}")
result = await wechat_service.poll_qr_status(qr_code_key, max_attempts)
if not result:
raise HTTPException(
status_code=500,
detail="轮询微信登录状态失败"
)
app_logger.info(f"轮询完成,结果: {result.success}")
return result
except HTTPException:
raise
except Exception as e:
app_logger.error(f"轮询微信登录状态时发生未知错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/wechat/callback",
summary="微信登录回调处理",
description="处理微信扫码后的回调,提取code参数")
async def wechat_callback(
code: str = Query(..., description="微信返回的授权码"),
state: str = Query(default="Lw==", description="状态参数")
):
"""
微信登录回调处理
这个接口用于演示微信回调的处理流程。
实际使用中,微信会重定向到配置的redirect_uri。
- **code**: 微信返回的授权码
- **state**: 状态参数
"""
try:
app_logger.info(f"收到微信回调,code: {code}, state: {state}")
# 自动进行登录验证
validate_response = await wechat_service.validate_wechat_login(code, state)
if validate_response and validate_response.success:
return {
"message": "微信登录成功",
"code": code,
"state": state,
"login_result": validate_response.dict(),
"next_step": "用户已成功登录,可以访问受保护的资源"
}
else:
return {
"message": "微信登录失败",
"code": code,
"state": state,
"error": validate_response.message if validate_response else "验证失败"
}
except Exception as e:
app_logger.error(f"处理微信回调时发生错误: {e}")
return {
"message": "处理微信回调失败",
"code": code,
"state": state,
"error": str(e)
}
@router.post("/password/login",
summary="密码登录",
description="使用手机号/邮箱,密码进行登录")
async def password_login(
login_request: PasswordLoginRequest
):
"""
密码登录滴答清单
- **username**: 登录账户(邮箱或手机号)
- **password**: 登录密码
返回登录结果,包含用户信息和认证令牌
"""
try:
app_logger.info(f"请求密码登录,用户名: {login_request.username}")
if not login_request.username or not login_request.password:
raise HTTPException(
status_code=400,
detail="用户名和密码不能为空"
)
login_response = await wechat_service.password_login(
login_request.username,
login_request.password
)
if not login_response:
return {"error": "密码登录失败,请稍后重试"}
# 记录日志
if 'errorCode' in login_response:
app_logger.info(f"密码登录失败,错误代码: {login_response.get('errorCode')}")
else:
user_id = login_response.get('userId', '')
app_logger.info(f"密码登录成功,用户ID: {user_id}")
# 直接返回原始响应,不进行任何封装
return login_response
except HTTPException:
raise
except Exception as e:
app_logger.error(f"密码登录时发生未知错误: {e}")
raise HTTPException(
status_code=500,
detail=f"服务器内部错误: {str(e)}"
)
@router.get("/health",
response_model=ApiResponse,
summary="健康检查",
description="检查认证服务的健康状态")
async def health_check() -> ApiResponse:
"""
健康检查接口
返回服务的健康状态信息
"""
return ApiResponse(
code=200,
message="认证服务运行正常",
data={
"service": "auth",
"status": "healthy",
"version": "1.0.0"
}
)
|
2977094657/DidaAPI | 1,193 | utils/logger.py | """日志配置模块"""
import sys
import os
from datetime import datetime
from loguru import logger
def setup_logger():
"""配置日志系统"""
# 移除默认的日志处理器
logger.remove()
# 使用默认配置
level = 'DEBUG'
format_str = '{time:YYYY-MM-DD HH:mm:ss} | {level} | {name}:{function}:{line} - {message}'
rotation = '1 day'
retention = '7 days'
# 控制台输出
logger.add(
sys.stdout,
format=format_str,
level=level,
colorize=True
)
# 获取当前日期,用于创建日志文件夹结构
now = datetime.now()
year = now.strftime('%Y')
month = now.strftime('%m')
day = now.strftime('%d')
# 创建日志文件夹结构: output/logs/年/月/日/
log_dir = f"output/logs/{year}/{month}/{day}"
os.makedirs(log_dir, exist_ok=True)
# 应用日志文件
logger.add(
f"{log_dir}/app.log",
format=format_str,
level=level,
rotation=rotation,
retention=retention,
encoding="utf-8"
)
# 错误日志单独文件
logger.add(
f"{log_dir}/error.log",
format=format_str,
level="ERROR",
rotation=rotation,
retention=retention,
encoding="utf-8"
)
return logger
# 初始化日志
app_logger = setup_logger()
|
281677160/openwrt-package | 17,188 | luci-app-passwall/root/usr/share/passwall/rule_update.lua | #!/usr/bin/lua
local api = require ("luci.passwall.api")
local name = api.appname
local uci = api.uci
local sys = api.sys
local jsonc = api.jsonc
local fs = api.fs
local arg1 = arg[1]
local arg2 = arg[2]
local arg3 = arg[3]
local nftable_name = "inet passwall"
local rule_path = "/usr/share/" .. name .. "/rules"
local reboot = 0
local gfwlist_update = "0"
local chnroute_update = "0"
local chnroute6_update = "0"
local chnlist_update = "0"
local geoip_update = "0"
local geosite_update = "0"
-- match comments/title/whitelist/ip address/excluded_domain
local comment_pattern = "^[#!\\[@]+"
local ip_pattern = "^%d+%.%d+%.%d+%.%d+"
local ip4_ipset_pattern = "^%d+%.%d+%.%d+%.%d+[%/][%d]+$"
local ip6_ipset_pattern = ":-[%x]+%:+[%x]-[%/][%d]+$"
local domain_pattern = "([%w%-]+%.[%w%.%-]+)[%/%*]*"
local excluded_domain = {"apple.com","sina.cn","sina.com.cn","baidu.com","byr.cn","jlike.com","weibo.com","zhongsou.com","youdao.com","sogou.com","so.com","soso.com","aliyun.com","taobao.com","jd.com","qq.com","bing.com"}
local gfwlist_url = uci:get(name, "@global_rules[0]", "gfwlist_url") or {"https://fastly.jsdelivr.net/gh/Loyalsoldier/v2ray-rules-dat@release/gfw.txt"}
local chnroute_url = uci:get(name, "@global_rules[0]", "chnroute_url") or {"https://ispip.clang.cn/all_cn.txt"}
local chnroute6_url = uci:get(name, "@global_rules[0]", "chnroute6_url") or {"https://ispip.clang.cn/all_cn_ipv6.txt"}
local chnlist_url = uci:get(name, "@global_rules[0]", "chnlist_url") or {"https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/accelerated-domains.china.conf","https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/apple.china.conf","https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/google.china.conf"}
local geoip_url = uci:get(name, "@global_rules[0]", "geoip_url") or "https://github.com/Loyalsoldier/geoip/releases/latest/download/geoip.dat"
local geosite_url = uci:get(name, "@global_rules[0]", "geosite_url") or "https://github.com/Loyalsoldier/v2ray-rules-dat/releases/latest/download/geosite.dat"
local asset_location = uci:get(name, "@global_rules[0]", "v2ray_location_asset") or "/usr/share/v2ray/"
local use_nft = uci:get(name, "@global_forwarding[0]", "use_nft") or "0"
local geo2rule = uci:get(name, "@global_rules[0]", "geo2rule") or "0"
local geoip_update_ok, geosite_update_ok = false, false
asset_location = asset_location:match("/$") and asset_location or (asset_location .. "/")
--兼容旧版本geo下载方式的配置,择机删除。
if geoip_url:match(".*/([^/]+)$") == "latest" then
geoip_url = "https://github.com/Loyalsoldier/geoip/releases/latest/download/geoip.dat"
end
if geosite_url:match(".*/([^/]+)$") == "latest" then
geosite_url = "https://github.com/Loyalsoldier/v2ray-rules-dat/releases/latest/download/geosite.dat"
end
if arg3 == "cron" then
arg2 = nil
end
local log = function(...)
if arg1 then
if arg1 == "log" then
api.log(...)
elseif arg1 == "print" then
local result = os.date("%Y-%m-%d %H:%M:%S: ") .. table.concat({...}, " ")
print(result)
end
end
end
local function gen_nftset(set_name, ip_type, tmp_file, input_file)
f = io.open(input_file, "r")
local element = f:read("*all")
f:close()
nft_file, err = io.open(tmp_file, "w")
nft_file:write('#!/usr/sbin/nft -f\n')
nft_file:write(string.format('define %s = {%s}\n', set_name, string.gsub(element, "%s*%c+", " timeout 3650d, ")))
if sys.call(string.format('nft "list set %s %s" >/dev/null 2>&1', nftable_name, set_name)) ~= 0 then
nft_file:write(string.format('add set %s %s { type %s; flags interval, timeout; timeout 2d; gc-interval 2d; auto-merge; }\n', nftable_name, set_name, ip_type))
end
nft_file:write(string.format('add element %s %s $%s\n', nftable_name, set_name, set_name))
nft_file:close()
sys.call(string.format('nft -f %s &>/dev/null',tmp_file))
os.remove(tmp_file)
end
--gen cache for nftset from file
local function gen_cache(set_name, ip_type, input_file, output_file)
local tmp_dir = "/tmp/"
local tmp_file = output_file .. "_tmp"
local tmp_set_name = set_name .. "_tmp"
gen_nftset(tmp_set_name, ip_type, tmp_file, input_file)
sys.call(string.format('nft list set %s %s | sed "s/%s/%s/g" | cat > %s', nftable_name, tmp_set_name, tmp_set_name, set_name, output_file))
sys.call(string.format('nft flush set %s %s', nftable_name, tmp_set_name))
sys.call(string.format('nft delete set %s %s', nftable_name, tmp_set_name))
end
-- curl
local function curl(url, file, valifile)
local args = {
"-skL", "-w %{http_code}", "--retry 3", "--connect-timeout 3", "--max-time 300", "--speed-limit 51200 --speed-time 15"
}
if file then
args[#args + 1] = "-o " .. file
end
if valifile then
args[#args + 1] = "--dump-header " .. valifile
end
local return_code, result = api.curl_auto(url, nil, args)
return tonumber(result)
end
--check excluded domain
local function check_excluded_domain(value)
for k,v in ipairs(excluded_domain) do
if value:find(v) then
return true
end
end
end
local function line_count(file_path)
local num = 0
for _ in io.lines(file_path) do
num = num + 1
end
return num;
end
local function non_file_check(file_path, vali_file)
if fs.readfile(file_path, 10) then
local size_str = sys.exec("grep -i 'Content-Length' " .. vali_file .. " | tail -n1 | sed 's/[^0-9]//g'")
local remote_file_size = tonumber(size_str ~= "" and size_str or nil)
local local_file_size = tonumber(fs.stat(file_path, "size"))
if remote_file_size and local_file_size then
if remote_file_size == local_file_size then
return nil;
else
log("下载文件大小校验出错,原始文件大小" .. remote_file_size .. "B,下载文件大小:" .. local_file_size .. "B。")
return true;
end
else
return nil;
end
else
log("下载文件读取出错。")
return true;
end
end
local function GeoToRule(rule_name, rule_type, out_path)
if not api.is_finded("geoview") then
log(rule_name .. "生成失败,缺少 geoview 组件。")
return false;
end
local geosite_path = asset_location .. "geosite.dat"
local geoip_path = asset_location .. "geoip.dat"
local file_path = (rule_type == "domain") and geosite_path or geoip_path
local arg
if rule_type == "domain" then
if rule_name == "gfwlist" then
arg = "-type geosite -list gfw"
else
arg = "-type geosite -list cn"
end
elseif rule_type == "ip4" then
arg = "-type geoip -list cn -ipv6=false"
elseif rule_type == "ip6" then
arg = "-type geoip -list cn -ipv4=false"
end
cmd = string.format("geoview -input '%s' %s -lowmem=true -output '%s'", file_path, arg, out_path)
sys.exec(cmd)
return true;
end
--fetch rule
local function fetch_rule(rule_name,rule_type,url,exclude_domain)
local sret = 200
local sret_tmp = 0
local domains = {}
local file_tmp = "/tmp/" ..rule_name.. "_tmp"
local vali_file = "/tmp/" ..rule_name.. "_vali"
local download_file_tmp = "/tmp/" ..rule_name.. "_dl"
local unsort_file_tmp = "/tmp/" ..rule_name.. "_unsort"
if geo2rule == "1" then
url = {"geo2rule"}
log(rule_name.. " 开始生成...")
else
log(rule_name.. " 开始更新...")
end
for k,v in ipairs(url) do
if v ~= "geo2rule" then
sret_tmp = curl(v, download_file_tmp..k, vali_file..k)
if sret_tmp == 200 and non_file_check(download_file_tmp..k, vali_file..k) then
log(rule_name.. " 第" ..k.. "条规则:" ..v.. "下载文件过程出错,尝试重新下载。")
os.remove(download_file_tmp..k)
os.remove(vali_file..k)
sret_tmp = curl(v, download_file_tmp..k, vali_file..k)
if sret_tmp == 200 and non_file_check(download_file_tmp..k, vali_file..k) then
sret = 0
sret_tmp = 0
log(rule_name.. " 第" ..k.. "条规则:" ..v.. "下载文件过程出错,请检查网络或下载链接后重试!")
end
end
else
if not GeoToRule(rule_name, rule_type, download_file_tmp..k) then return 1 end
sret_tmp = 200
end
if sret_tmp == 200 then
if rule_name == "gfwlist" and geo2rule == "0" then
local domains = {}
local gfwlist = io.open(download_file_tmp..k, "r")
local decode = api.base64Decode(gfwlist:read("*all"))
gfwlist:close()
gfwlist = io.open(download_file_tmp..k, "w")
gfwlist:write(decode)
gfwlist:close()
end
if rule_type == "domain" and exclude_domain == true then
for line in io.lines(download_file_tmp..k) do
line = line:gsub("full:", "")
if not (string.find(line, comment_pattern) or string.find(line, ip_pattern) or check_excluded_domain(line) or string.find(line, ":")) then
local match = string.match(line, domain_pattern)
if match then
domains[match] = true
end
end
end
elseif rule_type == "domain" then
for line in io.lines(download_file_tmp..k) do
line = line:gsub("full:", "")
if not (string.find(line, comment_pattern) or string.find(line, ip_pattern) or string.find(line, ":")) then
local match = string.match(line, domain_pattern)
if match then
domains[match] = true
end
end
end
elseif rule_type == "ip4" then
local out = io.open(unsort_file_tmp, "a")
for line in io.lines(download_file_tmp..k) do
if string.match(line, ip4_ipset_pattern) and not string.match(line, "^0%..*") then
out:write(string.format("%s\n", line))
end
end
out:close()
elseif rule_type == "ip6" then
local out = io.open(unsort_file_tmp, "a")
for line in io.lines(download_file_tmp..k) do
if string.match(line, ip6_ipset_pattern) and not string.match(line, "^::(/%d+)?$") then
out:write(string.format("%s\n", line))
end
end
out:close()
end
else
sret = 0
log(rule_name.. " 第" ..k.. "条规则:" ..v.. "下载失败,请检查网络或下载链接后重试!")
end
os.remove(download_file_tmp..k)
os.remove(vali_file..k)
end
if sret == 200 then
if rule_type == "domain" then
local out = io.open(unsort_file_tmp, "w")
for k,v in pairs(domains) do
out:write(string.format("%s\n", k))
end
out:close()
end
sys.call("LC_ALL=C sort -u " .. unsort_file_tmp .. " > " .. file_tmp)
os.remove(unsort_file_tmp)
local old_md5 = sys.exec("echo -n $(md5sum " .. rule_path .. "/" ..rule_name.. " | awk '{print $1}')"):gsub("\n", "")
local new_md5 = sys.exec("echo -n $([ -f '" ..file_tmp.. "' ] && md5sum " ..file_tmp.." | awk '{print $1}')"):gsub("\n", "")
if old_md5 ~= new_md5 then
local count = line_count(file_tmp)
if use_nft == "1" and (rule_type == "ip6" or rule_type == "ip4") then
local output_file = file_tmp.. ".nft"
if rule_type == "ip4" then
local set_name = "passwall_" ..rule_name
if rule_name == "chnroute" then
set_name = "passwall_chn"
end
gen_cache(set_name, "ipv4_addr", file_tmp, output_file)
elseif rule_type == "ip6" then
local set_name = "passwall_" ..rule_name
if rule_name == "chnroute6" then
set_name = "passwall_chn6"
end
gen_cache(set_name, "ipv6_addr", file_tmp, output_file)
end
sys.exec(string.format('mv -f %s %s', output_file, rule_path .. "/" ..rule_name.. ".nft"))
os.remove(output_file)
end
sys.exec("mv -f "..file_tmp .. " " ..rule_path .. "/" ..rule_name)
reboot = 1
log(rule_name.. " 更新成功,总规则数 " ..count.. " 条。")
else
log(rule_name.. " 版本一致,无需更新。")
end
else
log(rule_name.. " 文件下载失败!")
end
os.remove(file_tmp)
return 0
end
local function fetch_geofile(geo_name, geo_type, url)
local tmp_path = "/tmp/" .. geo_name
local asset_path = asset_location .. geo_name
local down_filename = url:match("^.*/([^/?#]+)")
local sha_url = url:gsub(down_filename, down_filename .. ".sha256sum")
local sha_path = tmp_path .. ".sha256sum"
local vali_file = tmp_path .. ".vali"
local function verify_sha256(sha_file)
return sys.call("sha256sum -c " .. sha_file .. " > /dev/null 2>&1") == 0
end
local sha_verify = curl(sha_url, sha_path) == 200
if sha_verify then
local f = io.open(sha_path, "r")
if f then
local content = f:read("*l")
f:close()
if content then
content = content:gsub(down_filename, tmp_path)
f = io.open(sha_path, "w")
if f then
f:write(content)
f:close()
end
end
end
if fs.access(asset_path) then
sys.call(string.format("cp -f %s %s", asset_path, tmp_path))
if verify_sha256(sha_path) then
log(geo_type .. " 版本一致,无需更新。")
return 0
end
end
end
local sret_tmp = curl(url, tmp_path, vali_file)
if sret_tmp == 200 and non_file_check(tmp_path, vali_file) then
log(geo_type .. " 下载文件过程出错,尝试重新下载。")
os.remove(tmp_path)
os.remove(vali_file)
sret_tmp = curl(url, tmp_path, vali_file)
if sret_tmp == 200 and non_file_check(tmp_path, vali_file) then
sret_tmp = 0
log(geo_type .. " 下载文件过程出错,请检查网络或下载链接后重试!")
end
end
if sret_tmp == 200 then
if sha_verify then
if verify_sha256(sha_path) then
sys.call(string.format("mkdir -p %s && cp -f %s %s", asset_location, tmp_path, asset_path))
reboot = 1
log(geo_type .. " 更新成功。")
if geo_type == "geoip" then
geoip_update_ok = true
else
geosite_update_ok = true
end
else
log(geo_type .. " 更新失败,请稍后重试或更换更新URL。")
return 1
end
else
if fs.access(asset_path) and sys.call(string.format("cmp -s %s %s", tmp_path, asset_path)) == 0 then
log(geo_type .. " 版本一致,无需更新。")
return 0
end
sys.call(string.format("mkdir -p %s && cp -f %s %s", asset_location, tmp_path, asset_path))
reboot = 1
log(geo_type .. " 更新成功。")
if geo_type == "geoip" then
geoip_update_ok = true
else
geosite_update_ok = true
end
end
else
log(geo_type .. " 更新失败,请稍后重试或更换更新URL。")
return 1
end
return 0
end
local function fetch_gfwlist()
fetch_rule("gfwlist","domain",gfwlist_url,true)
end
local function fetch_chnroute()
fetch_rule("chnroute","ip4",chnroute_url,false)
end
local function fetch_chnroute6()
fetch_rule("chnroute6","ip6",chnroute6_url,false)
end
local function fetch_chnlist()
fetch_rule("chnlist","domain",chnlist_url,false)
end
local function fetch_geoip()
fetch_geofile("geoip.dat","geoip",geoip_url)
end
local function fetch_geosite()
fetch_geofile("geosite.dat","geosite",geosite_url)
end
if arg2 then
string.gsub(arg2, '[^' .. "," .. ']+', function(w)
if w == "gfwlist" then
gfwlist_update = "1"
end
if w == "chnroute" then
chnroute_update = "1"
end
if w == "chnroute6" then
chnroute6_update = "1"
end
if w == "chnlist" then
chnlist_update = "1"
end
if w == "geoip" then
geoip_update = "1"
end
if w == "geosite" then
geosite_update = "1"
end
end)
else
gfwlist_update = uci:get(name, "@global_rules[0]", "gfwlist_update") or "1"
chnroute_update = uci:get(name, "@global_rules[0]", "chnroute_update") or "1"
chnroute6_update = uci:get(name, "@global_rules[0]", "chnroute6_update") or "1"
chnlist_update = uci:get(name, "@global_rules[0]", "chnlist_update") or "1"
geoip_update = uci:get(name, "@global_rules[0]", "geoip_update") or "1"
geosite_update = uci:get(name, "@global_rules[0]", "geosite_update") or "1"
end
if gfwlist_update == "0" and chnroute_update == "0" and chnroute6_update == "0" and chnlist_update == "0" and geoip_update == "0" and geosite_update == "0" then
os.exit(0)
end
log("开始更新规则...")
local function safe_call(func, err_msg)
xpcall(func, function(e)
log(e)
log(debug.traceback())
log(err_msg)
end)
end
local function remove_tmp_geofile(name)
os.remove("/tmp/" .. name .. ".dat")
os.remove("/tmp/" .. name .. ".dat.sha256sum")
os.remove("/tmp/" .. name .. ".dat.vali")
end
if geo2rule == "1" then
if geoip_update == "1" then
log("geoip 开始更新...")
safe_call(fetch_geoip, "更新geoip发生错误...")
remove_tmp_geofile("geoip")
end
if geosite_update == "1" then
log("geosite 开始更新...")
safe_call(fetch_geosite, "更新geosite发生错误...")
remove_tmp_geofile("geosite")
end
if geoip_update_ok then
safe_call(fetch_chnroute, "生成chnroute发生错误...")
safe_call(fetch_chnroute6, "生成chnroute6发生错误...")
end
if geosite_update_ok then
safe_call(fetch_gfwlist, "生成gfwlist发生错误...")
safe_call(fetch_chnlist, "生成chnlist发生错误...")
end
else
if gfwlist_update == "1" then
safe_call(fetch_gfwlist, "更新gfwlist发生错误...")
end
if chnroute_update == "1" then
safe_call(fetch_chnroute, "更新chnroute发生错误...")
end
if chnroute6_update == "1" then
safe_call(fetch_chnroute6, "更新chnroute6发生错误...")
end
if chnlist_update == "1" then
safe_call(fetch_chnlist, "更新chnlist发生错误...")
end
if geoip_update == "1" then
log("geoip 开始更新...")
safe_call(fetch_geoip, "更新geoip发生错误...")
remove_tmp_geofile("geoip")
end
if geosite_update == "1" then
log("geosite 开始更新...")
safe_call(fetch_geosite, "更新geosite发生错误...")
remove_tmp_geofile("geosite")
end
end
uci:set(name, "@global_rules[0]", "gfwlist_update", gfwlist_update)
uci:set(name, "@global_rules[0]", "chnroute_update", chnroute_update)
uci:set(name, "@global_rules[0]", "chnroute6_update", chnroute6_update)
uci:set(name, "@global_rules[0]", "chnlist_update", chnlist_update)
uci:set(name, "@global_rules[0]", "geoip_update", geoip_update)
uci:set(name, "@global_rules[0]", "geosite_update", geosite_update)
api.uci_save(uci, name, true)
if reboot == 1 then
if arg3 == "cron" then
if not fs.access("/var/lock/" .. name .. ".lock") then
sys.call("touch /tmp/lock/" .. name .. "_cron.lock")
end
end
log("重启服务,应用新的规则。")
uci:set(name, "@global[0]", "flush_set", "1")
api.uci_save(uci, name, true, true)
end
log("规则更新完毕...\n")
|
2977094657/DidaAPI | 2,107 | services/project_service.py | """项目管理服务模块"""
import httpx
from typing import Optional
from utils import app_logger
from core import urls
# 不再使用响应模型,直接返回原始响应
class ProjectService:
"""项目管理服务类"""
def __init__(self):
self.client = httpx.AsyncClient(timeout=30.0)
async def get_projects(self, auth_token: str, csrf_token: str) -> dict:
"""
获取项目/清单列表
Args:
auth_token: 认证令牌
csrf_token: CSRF令牌
Returns:
dict: 原始响应数据
"""
try:
# 构建请求URL
url = urls.build_dida_api_url(urls.DIDA_PROJECT_APIS["get_projects"])
# 构建请求头
headers = {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36',
'Accept': 'application/json, text/plain, */*',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8',
'Content-Type': 'application/json',
'X-Requested-With': 'XMLHttpRequest',
'X-Tz': 'Asia/Shanghai',
}
# 构建cookies
cookies = {
't': auth_token,
'_csrf_token': csrf_token
}
app_logger.info(f"请求获取项目列表: {url}")
# 发送请求
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
response_data = response.json()
app_logger.info(f"成功获取项目列表,项目数: {len(response_data) if isinstance(response_data, list) else 0}")
# 直接返回原始响应
return response_data
else:
app_logger.error(f"获取项目列表失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"获取项目列表时发生错误: {e}")
return {"error": str(e)}
async def close(self):
"""关闭HTTP客户端"""
await self.client.aclose()
# 全局项目服务实例
project_service = ProjectService()
|
2977094657/DidaAPI | 6,838 | services/pomodoro_service.py | """番茄专注服务模块"""
import httpx
from datetime import datetime, timezone, timedelta
from core import urls
class PomodoroService:
"""番茄专注服务类"""
def __init__(self):
self.client = httpx.AsyncClient(timeout=30.0)
def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证请求头"""
return {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36',
'Accept': 'application/json, text/plain, */*',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8',
'Content-Type': 'application/json',
'X-Requested-With': 'XMLHttpRequest',
'X-Tz': 'Asia/Shanghai',
}
def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证cookies"""
return {
't': auth_token,
'_csrf_token': csrf_token
}
def _convert_time_to_timestamp(self, time_str: str) -> int:
"""
将时间字符串转换为时间戳(毫秒)
Args:
time_str: 时间字符串,格式如 "2025-04-22T08:43:31.000+0000"
Returns:
int: 毫秒时间戳
"""
try:
# 解析时间字符串
dt = datetime.fromisoformat(time_str.replace('Z', '+00:00'))
# 转换为中国时间(UTC+8)
china_tz = timezone(timedelta(hours=8))
china_time = dt.astimezone(china_tz)
# 转换为时间戳(秒)然后转为毫秒
timestamp_ms = int(china_time.timestamp() * 1000)
return timestamp_ms
except Exception as e:
raise ValueError(f"时间转换失败: {e}")
async def get_general_for_desktop(self, auth_token: str, csrf_token: str) -> dict:
"""获取番茄专注概览(桌面版),直接返回原始响应"""
try:
url = urls.build_dida_api_url(urls.DIDA_POMODORO_APIS["general_for_desktop"])
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def get_focus_distribution(self, auth_token: str, csrf_token: str,
start_date: str, end_date: str) -> dict:
"""获取专注详情分布,直接返回原始响应"""
try:
endpoint = f"{urls.DIDA_POMODORO_APIS['focus_distribution']}/{start_date}/{end_date}"
url = urls.build_dida_api_url(endpoint)
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def get_focus_timeline(self, auth_token: str, csrf_token: str, to_timestamp: int = None) -> dict:
"""
获取专注记录时间线,支持分页
Args:
auth_token: 认证令牌
csrf_token: CSRF令牌
to_timestamp: 可选的时间戳参数,用于分页获取更早的数据
Returns:
dict: 原始API响应
"""
try:
url = urls.build_dida_api_url(urls.DIDA_POMODORO_APIS["focus_timeline"])
# 如果提供了时间戳参数,添加到URL中
if to_timestamp is not None:
url = f"{url}?to={to_timestamp}"
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def get_focus_heatmap(self, auth_token: str, csrf_token: str,
start_date: str, end_date: str) -> dict:
"""获取专注趋势热力图,直接返回原始响应"""
try:
endpoint = f"{urls.DIDA_POMODORO_APIS['focus_heatmap']}/{start_date}/{end_date}"
url = urls.build_dida_api_url(endpoint)
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def get_focus_time_distribution(self, auth_token: str, csrf_token: str,
start_date: str, end_date: str) -> dict:
"""获取专注时间分布(按时间段),直接返回原始响应"""
try:
endpoint = f"{urls.DIDA_POMODORO_APIS['focus_time_distribution']}/{start_date}/{end_date}"
url = urls.build_dida_api_url(endpoint)
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def get_focus_hour_distribution(self, auth_token: str, csrf_token: str,
start_date: str, end_date: str) -> dict:
"""获取专注时间按小时分布,直接返回原始响应"""
try:
endpoint = f"{urls.DIDA_POMODORO_APIS['focus_hour_distribution']}/{start_date}/{end_date}"
url = urls.build_dida_api_url(endpoint)
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def close(self):
"""关闭HTTP客户端"""
await self.client.aclose()
# 全局番茄专注服务实例
pomodoro_service = PomodoroService()
|
281677160/openwrt-package | 3,284 | luci-app-passwall/root/usr/share/passwall/test.sh | #!/bin/sh
CONFIG=passwall
LOG_FILE=/tmp/log/$CONFIG.log
echolog() {
local d="$(date "+%Y-%m-%d %H:%M:%S")"
#echo -e "$d: $1"
echo -e "$d: $1" >> $LOG_FILE
}
config_n_get() {
local ret=$(uci -q get "${CONFIG}.${1}.${2}" 2>/dev/null)
echo "${ret:=$3}"
}
lua_api() {
local func=${1}
[ -z "${func}" ] && {
echo "nil"
return
}
echo $(lua -e "local api = require 'luci.passwall.api' print(api.${func})")
}
test_url() {
local url=$1
local try=1
[ -n "$2" ] && try=$2
local timeout=2
[ -n "$3" ] && timeout=$3
local extra_params=$4
curl --help all | grep "\-\-retry-all-errors" > /dev/null
[ $? == 0 ] && extra_params="--retry-all-errors ${extra_params}"
status=$(/usr/bin/curl -I -o /dev/null -skL $extra_params --connect-timeout ${timeout} --retry ${try} -w %{http_code} "$url")
case "$status" in
204|\
200)
status=200
;;
esac
echo $status
}
test_proxy() {
result=0
status=$(test_url "https://www.google.com/generate_204" ${retry_num} ${connect_timeout})
if [ "$status" = "200" ]; then
result=0
else
status2=$(test_url "https://www.baidu.com" ${retry_num} ${connect_timeout})
if [ "$status2" = "200" ]; then
result=1
else
result=2
ping -c 3 -W 1 223.5.5.5 > /dev/null 2>&1
[ $? -eq 0 ] && {
result=1
}
fi
fi
echo $result
}
url_test_node() {
result=0
local node_id=$1
local _type=$(echo $(config_n_get ${node_id} type) | tr 'A-Z' 'a-z')
[ -n "${_type}" ] && {
if [ "${_type}" == "socks" ]; then
local _address=$(config_n_get ${node_id} address)
local _port=$(config_n_get ${node_id} port)
[ -n "${_address}" ] && [ -n "${_port}" ] && {
local curlx="socks5h://${_address}:${_port}"
local _username=$(config_n_get ${node_id} username)
local _password=$(config_n_get ${node_id} password)
[ -n "${_username}" ] && [ -n "${_password}" ] && curlx="socks5h://${_username}:${_password}@${_address}:${_port}"
}
else
local _tmp_port=$(/usr/share/${CONFIG}/app.sh get_new_port 61080 tcp)
/usr/share/${CONFIG}/app.sh run_socks flag="url_test_${node_id}" node=${node_id} bind=127.0.0.1 socks_port=${_tmp_port} config_file=url_test_${node_id}.json
local curlx="socks5h://127.0.0.1:${_tmp_port}"
fi
sleep 1s
# 兼容 curl 8.6 time_starttransfer 错误
local _cmd="-V 2>/dev/null | head -n 1 | awk '{print \$2}' | cut -d. -f1,2 | tr -d ' \\n'"
local _curl="/usr/bin/curl"
local curl_ver=$(lua_api "get_bin_version_cache(\"${_curl}\", \"${_cmd}\")")
local curl_arg="-w %{http_code}:%{time_starttransfer} http://"
[ "${curl_ver}" = "8.6" ] && curl_arg="-w %{http_code}:%{time_appconnect} https://"
local chn_list=$(config_n_get @global[0] chn_list direct)
local probeUrl="www.google.com/generate_204"
[ "${chn_list}" = "proxy" ] && probeUrl="www.baidu.com"
result=$(${_curl} --max-time 5 -o /dev/null -I -skL -x ${curlx} ${curl_arg}${probeUrl})
# 结束 SS 插件进程
local pid_file="/tmp/etc/${CONFIG}/url_test_${node_id}_plugin.pid"
[ -s "$pid_file" ] && kill -9 "$(head -n 1 "$pid_file")" >/dev/null 2>&1
pgrep -af "url_test_${node_id}" | awk '! /test\.sh/{print $1}' | xargs kill -9 >/dev/null 2>&1
rm -rf /tmp/etc/${CONFIG}/*url_test_${node_id}*.*
}
echo $result
}
arg1=$1
shift
case $arg1 in
test_url)
test_url $@
;;
url_test_node)
url_test_node $@
;;
esac
|
294coder/Efficient-MIF | 4,574 | readmes/FeINFN.md | # Fourier-enhanced Implicit Neural Fusion Network for Multispectral and Hyperspectral Image Fusion
<div align="center">
<p style="text-align: center">
<a href="https://scholar.google.com/citations?user=E5KO9XsAAAAJ&hl=en", style="font-size: 18px;">Yu-Jie Liang</a>
 
<a href="https://scholar.google.com/citations?user=pv61p_EAAAAJ&hl=en", style="font-size: 18px;">Zihan Cao</a>
 
<a href="https://scholar.google.com/citations?user=JZag1WIAAAAJ&hl=en", style="font-size: 18px;"> Shangqi Deng </a>
 
<a style="font-size: 18px;"> Hong-Xia Dou </a>
 
<a href="https://liangjiandeng.github.io/", style="font-size: 18px;"> Liang-Jian Deng </a>
<br>
<a style="font-size: 16px;"> University of Electronic Science & Technology of China </a>
</p>
</div>
[](https://arxiv.org/abs/2404.15174)
abtract: Recently, implicit neural representations (INR) have made significant strides in various vision-related domains, providing a novel solution for Multispectral and Hyperspectral Image Fusion (MHIF) tasks. However, INR is prone to losing high-frequency information and is confined to the lack of global perceptual capabilities. To address these issues, this paper introduces a Fourier-enhanced Implicit Neural Fusion Network (FeINFN) specifically designed for MHIF task, targeting the following phenomena: \textit{The Fourier amplitudes of the HR-HSI latent code and LR-HSI are remarkably similar; however, their phases exhibit different patterns.} In FeINFN, we innovatively propose a spatial and frequency implicit fusion function (Spa-Fre IFF), helping INR capture high-frequency information and expanding the receptive field. Besides, a new decoder employing a complex Gabor wavelet activation function, called Spatial-Frequency Interactive Decoder (SFID), is invented to enhance the interaction of INR features. Especially, we further theoretically prove that the Gabor wavelet activation possesses a time-frequency tightness property that favors learning the optimal bandwidths in the decoder. Experiments on two benchmark MHIF datasets verify the state-of-the-art (SOTA) performance of the proposed method, both visually and quantitatively. Also, ablation studies demonstrate the mentioned contributions.
<html>
<body>
<div class="image-container" style="text-align: center;">
<img src="../teasers/FeINFN.png" alt="Image 1" width="100%">
</div>
</body>
</html>
# Model
We implement FeINFN with Pytorch and you can find it at [`model/FeINFN.py`](../model/FeINFN.py).
## Traning
To train the model, running the following commands:
```shell
CUDA_VISIBLE_DEVICES="0" \
NCCL_P2P_LEVEL="NVL" \
NCCL_P2P_DISABLE="1" \
NCCL_IB_DISABLE="1" \
OMP_NUM_THREADS="6" \
accelerate launch \
--config_file configs/huggingface/accelerate.yaml \
accelerate_main.py \
--proj_name FeINFN \
--arch FeINFN \
--dataset <dataset_name> \
--num_worker 0 -e 2000 -b 4 --aug_probs 0. 0. --loss l1ssim --grad_accum_steps 2 \
--checkpoint_every_n 20 --val_n_epoch 20 \
--comment "FeINFN config on <dataset_name> dataset model" \
--log_metric \
--logger_on \
```
> check the `model/__init__.py` if the FeINFN network is not registered.
Checkpoints, running, and Tensorboard logs will be saved at `log_file/`.
## Testing
You can refer to the testing script [`torch_inference_on_sharpening.py`](../torch_inference_on_sharpening.py) to test the model.
To test the metrics, please see the main guidance in [`README.md`](../README.md).
CAVE pretrained model is released [here](https://pan.baidu.com/s/1JMRidVmoXZ6tKbtsGmG__w?pwd=mtef) (BaiduYun Disk).
For sharpening tasks (including pansharpening and HMIF tasks), you simply test the metrics in Matlab:
``` matlab
cd Pansharpening_Hyper_SR_Matlab_Test_Package
%% when testing the reduced-resolution metrics on MHIF tasks
% Args:
% path: the saved fused image `.mat` file, find it in `visualized_img/`
% ratio: upscale ratio, e.g., 4
% full_res: we keep it to 0, not changed
% const: max value of the dataset (CAVE x4: 1, Harvard x4: 1, CAVE x8: 1, Harvard x8: 1)
analysis_ref_batched_images(path, ratio, full_res, const)
```
# Citation
If you find this work useful, please consider citing:
```bibtex
@article{liang2024fourier,
title={Fourier-enhanced Implicit Neural Fusion Network for Multispectral and Hyperspectral Image Fusion},
author={Liang, Yu-Jie and Cao, Zihan and Deng, ShangQi and Dou Hong-Xia and Deng, Liang-Jian},
journal={arXiv preprint arXiv:2404.15174},
year={2024}
}
```
|
294coder/Efficient-MIF | 3,890 | readmes/lformer.md | # Linearly-evolved Transformer for Pan-sharpening
<div align="center">
<p style="text-align: center">
<a style="font-size: 18px;"> JunMing Hou* </a>
 
<a href="https://scholar.google.com/citations?user=pv61p_EAAAAJ&hl=en", style="font-size: 18px;">Zihan Cao*</a>
 
<a style="font-size: 18px;"> Naishan Zheng </a>
 
<a style="font-size: 18px;"> Xuan Li </a>
 
<a style="font-size: 18px;"> Xiaoyu Chen </a>
 
<br>
<a style="font-size: 18px;"> XinYang Li </a>
 
<a style="font-size: 18px;"> Xiaofeng Cong</a>
 
<a style="font-size: 18px;"> Man Zhou </a>
 
<a style="font-size: 18px;"> Danfeng Hong </a>
 
<br>
<a style="font-size: 16px;"> University of Electronic Science Technology of China </a>
<br>
<a style="font-size: 16px;"> Southeast University </a>
<br>
<a style="font-size: 16px;"> University of Science and Technology </a>
</p>
</div>
[](https://arxiv.org/abs/2404.12804)
# Fast testing
We provide [pretrained weights](https://pan.baidu.com/s/1keK5eAIrZcPPgoEr8bcW5A?pwd=y2t9) and a fast testing script to test the performance of our model.
To run the testing script, please refer to `torch_inference_on_sharpening.py` and adapt following steps:
1. modify the `path` for datset;
2. change the `dataset_type`;
3. set `full_res` to the full resolution or reduced resolution of datasets (for pansharpening).
4. change the yaml file for configurate the model. For LFormer, it's in `configs/lformer_config.yaml`.
```yaml
network_configs:
lformer:
pan_dim: 1
lms_dim: 4 # 4 for GF2, 8 for WV3, 31 for CAVE x4
attn_dim: 64
hp_dim: 64
n_stage: 5
patch_merge: yes
crop_batch_size: 64
patch_size_list: [16, 64, 64]
scale: 4
```
To test the metrics, please see the main guidance in [`README.md`](../README.md).
For sharpening tasks (including pansharpening and HMIF tasks), you simply test the metrics in Matlab:
``` matlab
cd Pansharpening_Hyper_SR_Matlab_Test_Package
%% when testing the reduced-resolution metrics on pansharpening or HMIF tasks
% Args:
% path: the saved fused image `.mat` file, find it in `visualized_img/`
% ratio: upscale ratio, e.g., 4
% full_res: we keep it to 1, not changed
% const: max value of the dataset (WV3: 2047, GF2: 1023, CAVE x8: 1, Harvard x8: 1)
analysis_ref_batched_images(path, ratio, full_res, const)
%% when testing full-resolution metrics on pansharpening datasets
% Args:
% path: the saved fused image `.mat` file, find it in `visualized_img/`
% ratio: upscale ratio, e.g., 4
% sensor: the sensor name ('WV3', 'GF2', 'CAVE', 'Harvard')
analysis_unref_batched_images(path, ratio, sensor)
```
# Train
You can train the LFormer model by run the commands:
```shell
CUDA_VISIBLE_DEVICES="0" \
NCCL_P2P_LEVEL="NVL" \
NCCL_P2P_DISABLE="1" \
NCCL_IB_DISABLE="1" \
OMP_NUM_THREADS="6" \
accelerate launch \
--config_file configs/huggingface/accelerate.yaml \
accelerate_main.py \
--proj_name LFormer \
--arch LFormer \
--dataset <dataset_name> \
--num_worker 6 -e 800 -b 32 --aug_probs 0. 0. --loss l1ssim --grad_accum_steps 2 \
--checkpoint_every_n 20 --val_n_epoch 20 \
--comment "LFormer config on wv3 dataset model" \
--log_metric \
--logger_on \
```
> Other DDP training (or using deepspeed, please see Huggingface Accelerate documenation).
# Citation
If you find this code useful for your research, please consider citing:
```
@inproceedings{
hou2024linearlyevolved,
title={Linearly-evolved Transformer for Pan-sharpening},
author={Junming Hou and Zihan Cao and Naishan Zheng and Xuan Li and Xiaoyu Chen and Xinyang Liu and Xiaofeng Cong and Danfeng Hong and Man Zhou},
booktitle={ACM Multimedia 2024},
year={2024},
url={https://openreview.net/forum?id=pCxZTmGr4O}
}
``` |
294coder/Efficient-MIF | 5,435 | readmes/le-mamba.md | # A Novel State Space Model with Local Enhancement and State Sharing for Image Fusion
<div align="center">
<p style="text-align: center">
<a href="https://scholar.google.com/citations?user=pv61p_EAAAAJ&hl=en", style="font-size: 18px;">Zihan Cao</a>
 
<a href="https://xiaoxiao-woo.github.io/", style="font-size: 18px;"> Xiao Wu </a>
 
<a href="https://liangjiandeng.github.io/", style="font-size: 18px;"> Liang-Jian Deng </a>
<a style="font-size: 18px;"> Yu Zhong </a>
<br>
<a style="font-size: 16px;"> University of Electronic Science Technology of China </a>
</p>
</div>
[](https://arxiv.org/abs/2404.09293)
abtract: In image fusion tasks, images from different sources possess distinct characteristics. This has driven the development of numerous methods to explore better ways of fusing them while preserving their respective characteristics.
Mamba, as a state space model, has emerged in the field of natural language processing. Recently, many studies have attempted to extend Mamba to vision tasks. However, due to the nature of images different from causal language sequences, the limited state capacity of Mamba weakens its ability to model image information. Additionally, the sequence modeling ability of Mamba is only capable of spatial information and cannot effectively capture the rich spectral information in images. Motivated by these challenges, we customize and improve the vision Mamba network designed for the image fusion task. Specifically, we propose the local-enhanced vision Mamba block, dubbed as LEVM. The LEVM block can improve local information perception of the network and simultaneously learn local and global spatial information. Furthermore, we propose the state sharing technique to enhance spatial details and integrate spatial and spectral information. Finally, the overall network is a multi-scale structure based on vision Mamba, called LE-Mamba. Extensive experiments show the proposed methods achieve state-of-the-art results on multispectral pansharpening and multispectral and hyperspectral image fusion datasets, and demonstrate the effectiveness of the proposed approach.
<html>
<body>
<div class="image-container" style="text-align: center;">
<img src="figs/le-mamba-teaser.png" alt="Image 1" width="100%">
</div>
</body>
</html>
# Model
We implement LE-Mamba with Pytorch and you can find it at [`model/LEMamba.py`](../model/LEMamba.py).
## Traning
To train the model, you should first compile the `select_scan` operator by running the following command:
```shell
cd model/module/mamba_scan
pip install -e .
```
Then, you should run the following command to test the model if it is correctly installed:
```shell
pytest model/test/LEMamba_test.py
```
If the test is passed, you can start training the model by running the following commands:
```shell
CUDA_VISIBLE_DEVICES="0" \
NCCL_P2P_LEVEL="NVL" \
NCCL_P2P_DISABLE="1" \
NCCL_IB_DISABLE="1" \
OMP_NUM_THREADS="6" \
accelerate launch \
--config_file configs/huggingface/accelerate.yaml \
accelerate_main.py \
--proj_name LE-Mamba \
--arch LEMamba \
--dataset <dataset_name> \
--num_worker 6 -e 800 -b 32 --aug_probs 0. 0. --loss l1ssim --grad_accum_steps 2 \
--checkpoint_every_n 20 --val_n_epoch 20 \
--comment "LE-Mamba config on wv3 dataset model" \
--log_metric \
--logger_on \
```
> check the `model/__init__.py` if the LE-Mamba network is not registered.
Checkpoints, running, and Tensorboard logs will be saved at `log_file/`.
## Testing
You can refer to the testing script [`torch_inference_on_sharpening.py`](../torch_inference_on_sharpening.py) to test the model.
To test the metrics, please see the main guidance in [`README.md`](../README.md).
For sharpening tasks (including pansharpening and HMIF tasks), you simply test the metrics in Matlab:
``` matlab
cd Pansharpening_Hyper_SR_Matlab_Test_Package
%% when testing the reduced-resolution metrics on pansharpening or HMIF tasks
% Args:
% path: the saved fused image `.mat` file, find it in `visualized_img/`
% ratio: upscale ratio, e.g., 4
% full_res: we keep it to 1, not changed
% const: max value of the dataset (WV3: 2047, GF2: 1023, CAVE x8: 1, Harvard x8: 1)
analysis_ref_batched_images(path, ratio, full_res, const)
%% when testing full-resolution metrics on pansharpening datasets
% Args:
% path: the saved fused image `.mat` file, find it in `visualized_img/`
% ratio: upscale ratio, e.g., 4
% sensor: the sensor name ('WV3', 'GF2', 'CAVE', 'Harvard')
analysis_unref_batched_images(path, ratio, sensor)
```
<!-- # Performances
LE-Mamba reaches SOTA performances on widely-used Pansharpening and HMIF datasets. Here are some metrics:
<html>
<body>
<div class="image-container" style="text-align: center;">
<img src="figs/le-mamba-pansharpening.png" alt="Image 1" width="100%">
</div>
<div class="image-container" style="text-align: center;">
<img src="figs/le-mamba-HMIF.png" alt="Image 1" width="100%">
</div>
</body>
</html> -->
# Citation
If you find this work useful, please consider citing:
```bibtex
@inproceedings{
cao2024novel,
title={A novel state space model with local enhancement and state sharing for image fusion},
author={Cao, Zihan and Wu, Xiao and Deng, Liang-Jian and Zhong, Yu},
booktitle={ACM Multimedia 2024 (ACM MM)},
year={2024}
}
``` |
2977094657/DidaAPI | 10,585 | services/dida_service.py | """滴答清单API服务模块"""
import uuid
import time
from typing import Optional, Dict, Any, List
import httpx
from utils import app_logger
from core import config, db, urls
from models import TasksResponse, TaskItem
class DidaAPIService:
"""滴答清单API服务类"""
def __init__(self):
self.request_config = config.get('request_config', {})
self.client = httpx.AsyncClient(timeout=self.request_config.get('timeout', 30.0))
# 从数据库获取当前活跃的认证会话
self.current_session = None
self._load_active_session()
def _load_active_session(self):
"""从数据库加载活跃的认证会话"""
try:
# 从数据库获取最新的活跃会话
session_data = db.get_latest_active_session()
if session_data:
self.current_session = {
'session_id': session_data['session_id'],
'auth_token': session_data['token'],
'csrf_token': session_data['csrf_token'],
'is_active': session_data['is_active']
}
app_logger.info(f"已从数据库恢复认证会话: {session_data['session_id']}")
else:
app_logger.info("数据库中没有找到活跃的认证会话")
except Exception as e:
app_logger.error(f"加载认证会话失败: {e}")
def set_auth_session(self, auth_token: str, csrf_token: str) -> str:
"""设置认证会话"""
session_id = str(uuid.uuid4())
self.current_session = {
'session_id': session_id,
'auth_token': auth_token,
'csrf_token': csrf_token,
'is_active': True
}
# 保存到数据库
db.save_user_session({
'session_id': session_id,
'token': auth_token,
'csrf_token': csrf_token,
'is_active': True
})
app_logger.info(f"设置认证会话成功: {session_id}")
return session_id
def get_session_status(self) -> Dict[str, Any]:
"""获取当前会话状态"""
if self.current_session:
return {
"has_session": True,
"session_id": self.current_session.get('session_id'),
"is_active": self.current_session.get('is_active', False)
}
else:
return {
"has_session": False,
"session_id": None,
"is_active": False
}
def _get_auth_headers(self) -> Dict[str, str]:
"""获取认证请求头"""
if not self.current_session:
raise ValueError("未设置认证会话,请先登录")
# 生成traceid
traceid = f"{int(time.time() * 1000):x}{uuid.uuid4().hex[:8]}"
headers = {
'Accept': 'application/json, text/plain, */*',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7',
'Cache-Control': 'no-cache',
'Origin': 'https://dida365.com',
'Pragma': 'no-cache',
'Referer': 'https://dida365.com/',
'Sec-Ch-Ua': '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
'Sec-Ch-Ua-Mobile': '?0',
'Sec-Ch-Ua-Platform': '"Windows"',
'Sec-Fetch-Dest': 'empty',
'Sec-Fetch-Mode': 'cors',
'Sec-Fetch-Site': 'same-site',
'User-Agent': self.request_config.get('user_agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'),
'X-Csrftoken': self.current_session['csrf_token'],
'X-Device': self.request_config.get('device_info', '{}'),
'Hl': self.request_config.get('language', 'zh_CN'),
'X-Tz': self.request_config.get('timezone', 'Asia/Shanghai'),
'Traceid': traceid
}
return headers
def _get_auth_cookies(self) -> Dict[str, str]:
"""获取认证cookies"""
if not self.current_session:
raise ValueError("未设置认证会话,请先登录")
cookies = {
't': self.current_session['auth_token'],
'_csrf_token': self.current_session['csrf_token']
}
return cookies
async def get_all_tasks(self) -> dict:
"""
获取所有任务
Returns:
dict: 原始响应数据
"""
try:
if not self.current_session:
return {"error": "no_auth_session", "message": "未设置认证会话,请先登录"}
# 使用统一的URL构建函数
url = urls.build_dida_api_url(urls.DIDA_TASK_APIS["get_all_tasks"])
# 获取认证头和cookies
headers = self._get_auth_headers()
cookies = self._get_auth_cookies()
app_logger.info(f"请求获取所有任务: {url}")
app_logger.debug(f"请求头: {headers}")
# 发送请求
response = await self.client.get(url, headers=headers, cookies=cookies)
# 记录响应信息
app_logger.info(f"任务响应状态码: {response.status_code}")
app_logger.debug(f"任务响应头: {dict(response.headers)}")
if response.status_code == 200:
# 解析响应数据
response_data = response.json()
app_logger.info(f"成功获取任务数据,响应长度: {len(str(response_data))}")
app_logger.debug(f"任务响应数据: {response_data}")
# 直接返回原始响应
return response_data
else:
app_logger.error(f"获取任务失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"获取任务时发生错误: {e}")
return {"error": str(e)}
async def get_completed_tasks(self, to: Optional[str] = None, status: str = "Completed") -> dict:
"""
获取已完成或已放弃的任务(支持分页)
Args:
to: 分页参数,使用上次响应最后一个任务的completedTime字段
如果为None,则获取第一页
后续请求使用上次响应最后一个任务的completedTime字段(原始格式)
原始格式:2025-03-15T13:30:54.000+0000
API格式:2025-03-15 13:30:54
status: 任务状态,支持以下值:
- "Completed": 已完成的任务
- "Abandoned": 已放弃的任务
Returns:
dict: 原始响应数据,包含任务列表
Note:
分页机制:
- 第一次请求:不传to参数
- 后续请求:使用上次响应最后一个任务的completedTime字段作为to参数
- completedTime原始格式:2025-03-15T13:30:54.000+0000
- API需要格式:2025-03-15 13:30:54
- URL示例:https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Completed
"""
try:
if not self.current_session:
return {"error": "no_auth_session", "message": "未设置认证会话,请先登录"}
# 构建URL
base_url = urls.build_dida_api_url(urls.DIDA_TASK_APIS["get_completed_tasks"])
# 构建查询参数
params = {
"from": "", # 固定为空
"status": status # 支持Completed或Abandoned
}
# 如果提供了to参数,则添加到查询参数中
if to:
# 将completedTime格式转换为滴答清单API需要的格式
# 从 2025-03-15T13:30:54.000+0000 转换为 2025-03-15 13:30:54
formatted_to = to.replace('T', ' ').replace('.000+0000', '')
params["to"] = formatted_to
# 第一次请求不添加to参数
# 获取认证头和cookies
headers = self._get_auth_headers()
cookies = self._get_auth_cookies()
app_logger.info(f"请求获取已完成任务: {base_url}")
app_logger.info(f"查询参数: {params}")
app_logger.debug(f"请求头: {headers}")
# 发送请求
response = await self.client.get(base_url, headers=headers, cookies=cookies, params=params)
# 记录响应信息
app_logger.info(f"已完成任务响应状态码: {response.status_code}")
app_logger.debug(f"已完成任务响应头: {dict(response.headers)}")
if response.status_code == 200:
# 解析响应数据
response_data = response.json()
task_count = len(response_data) if isinstance(response_data, list) else 0
app_logger.info(f"成功获取已完成任务数据,任务数量: {task_count}")
app_logger.debug(f"已完成任务响应数据: {response_data}")
# 直接返回原始响应
return response_data
else:
app_logger.error(f"获取已完成任务失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"获取已完成任务时发生错误: {e}")
return {"error": str(e)}
async def get_trash_tasks(self, limit: int = 50, task_type: int = 1) -> dict:
"""
获取垃圾桶中的任务
Args:
limit: 每页任务数量,默认50
task_type: 任务类型,默认1
Returns:
dict: 原始响应数据,包含垃圾桶任务列表
Note:
响应格式:
{
"tasks": [...], # 任务列表
"next": 0 # 下一页标识
}
"""
try:
if not self.current_session:
return {"error": "no_auth_session", "message": "未设置认证会话,请先登录"}
# 构建URL
base_url = urls.build_dida_api_url(urls.DIDA_TASK_APIS["get_trash_tasks"])
# 构建查询参数
params = {
"limit": limit,
"type": task_type
}
# 获取认证头和cookies
headers = self._get_auth_headers()
cookies = self._get_auth_cookies()
app_logger.info(f"请求获取垃圾桶任务: {base_url}")
app_logger.info(f"查询参数: {params}")
app_logger.debug(f"请求头: {headers}")
# 发送请求
response = await self.client.get(base_url, headers=headers, cookies=cookies, params=params)
# 记录响应信息
app_logger.info(f"垃圾桶任务响应状态码: {response.status_code}")
app_logger.debug(f"垃圾桶任务响应头: {dict(response.headers)}")
if response.status_code == 200:
# 解析响应数据
response_data = response.json()
task_count = len(response_data.get('tasks', [])) if isinstance(response_data, dict) else 0
app_logger.info(f"成功获取垃圾桶任务数据,任务数量: {task_count}")
app_logger.debug(f"垃圾桶任务响应数据: {response_data}")
# 直接返回原始响应
return response_data
else:
app_logger.error(f"获取垃圾桶任务失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"获取垃圾桶任务时发生错误: {e}")
return {"error": str(e)}
async def close(self):
"""关闭HTTP客户端"""
await self.client.aclose()
# 全局滴答清单API服务实例
dida_service = DidaAPIService()
|
2977094657/DidaAPI | 3,363 | services/statistics_service.py | """统计服务模块"""
import httpx
from utils import app_logger
from core import urls
class StatisticsService:
"""统计服务类"""
def __init__(self):
self.client = httpx.AsyncClient(timeout=30.0)
def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证请求头"""
return {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36',
'Accept': 'application/json, text/plain, */*',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8',
'Content-Type': 'application/json',
'X-Requested-With': 'XMLHttpRequest',
'X-Tz': 'Asia/Shanghai',
}
def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证cookies"""
return {
't': auth_token,
'_csrf_token': csrf_token
}
async def get_user_ranking(self, auth_token: str, csrf_token: str) -> dict:
"""获取用户排名统计,直接返回原始响应"""
try:
url = urls.build_dida_api_url(urls.DIDA_STATISTICS_APIS["user_ranking"]).replace('/v2/', '/v3/')
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def get_general_statistics(self, auth_token: str, csrf_token: str) -> dict:
"""获取通用统计信息,直接返回原始响应"""
try:
url = urls.build_dida_api_url(urls.DIDA_STATISTICS_APIS["general_statistics"])
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def get_task_statistics(self, auth_token: str, csrf_token: str,
start_date: str, end_date: str) -> dict:
"""获取任务统计信息,直接返回原始响应"""
try:
endpoint = f"{urls.DIDA_STATISTICS_APIS['task_statistics']}/{start_date}/{end_date}"
url = urls.build_dida_api_url(endpoint)
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
return response.json()
else:
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
return {"error": str(e)}
async def close(self):
"""关闭HTTP客户端"""
await self.client.aclose()
# 全局统计服务实例
statistics_service = StatisticsService()
|
294coder/Efficient-MIF | 1,073 | Pansharpening_Hyper_SR_Matlab_Test_Package/quality_assess.m | function [mean_psnr, mean_ssim] = quality_assess(imagery1, imagery2)
%==========================================================================
% Evaluates the quality assessment indices for two tensors.
%
% Syntax:
% [mpsnr, mssim] = quality_access(imagery1, imagery2)
%
% Input:
% imagery1 - the reference tensor
% imagery2 - the target tensor
% NOTE: the tensor is a I1*I2*...*IN array and DYNAMIC RANGE [0, 255].
% Output:
% mpsnr - Peak Signal-to-Noise Ratio
% mssim - Structure SIMilarity
%==========================================================================
Nway = size(imagery1);
if length(Nway)>3
imagery1 = reshape(imagery1,Nway(1),Nway(2),[]);
imagery2 = reshape(imagery2,Nway(1),Nway(2),[]);
end
psnr = zeros(prod(Nway(3:end)),1);
ssim = psnr;
for ii = 1:prod(Nway(3:end))
psnr(ii) = psnr_index(imagery1(:, :, ii), imagery2(:, :, ii));
% 因为ssim_index是邓尚琦给的代码,需要乘255
ssim(ii) = ssim_index(imagery1(:, :, ii)*255, imagery2(:, :, ii)*255);
end
mean_psnr = mean(psnr);
mean_ssim = mean(ssim);
%out = [mean(psnr), mean(ssim)];
|
2977094657/DidaAPI | 2,429 | services/user_service.py | """用户信息服务模块"""
import httpx
from core import urls
from utils import app_logger
class UserService:
"""用户信息服务类"""
def __init__(self):
self.client = httpx.AsyncClient(timeout=30.0)
def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证headers"""
return {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36',
'Accept': 'application/json, text/plain, */*',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8',
'Content-Type': 'application/json;charset=UTF-8',
'X-Requested-With': 'XMLHttpRequest',
'x-csrftoken': csrf_token,
'Referer': 'https://dida365.com/',
'Origin': 'https://dida365.com'
}
def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证cookies"""
return {
't': auth_token,
'_csrf_token': csrf_token
}
async def get_user_profile(self, auth_token: str, csrf_token: str) -> dict:
"""
获取用户信息,直接返回原始响应
Args:
auth_token: 认证令牌
csrf_token: CSRF令牌
Returns:
dict: 原始API响应
"""
try:
url = urls.build_dida_api_url(urls.DIDA_AUTH_APIS["user_profile"])
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
app_logger.info(f"请求获取用户信息: {url}")
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
response_data = response.json()
app_logger.info(f"成功获取用户信息,用户名: {response_data.get('username', 'N/A')}")
# 直接返回原始响应
return response_data
else:
app_logger.error(f"获取用户信息失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"获取用户信息时发生错误: {e}")
return {"error": str(e)}
async def close(self):
"""关闭HTTP客户端"""
await self.client.aclose()
# 全局用户服务实例
user_service = UserService()
|
2977094657/DidaAPI | 23,614 | services/export_service.py | """任务导出服务"""
import io
from typing import Dict, List, Any, Optional
from datetime import datetime
import pandas as pd
from utils import app_logger
from services.dida_service import dida_service
from services.pomodoro_service import pomodoro_service
from core import urls
class ExportService:
"""任务导出服务类"""
def __init__(self):
self.dida_service = dida_service
async def export_tasks_to_excel(self) -> Dict[str, Any]:
"""
导出所有任务到Excel文件
Returns:
dict: 包含Excel文件内容和元数据的响应
"""
try:
app_logger.info("开始导出任务到Excel")
# 获取所有任务数据
all_tasks_data = await self._get_all_tasks_data()
completed_tasks_data = await self._get_completed_tasks_data()
abandoned_tasks_data = await self._get_abandoned_tasks_data()
trash_tasks_data = await self._get_trash_tasks_data()
if not all_tasks_data and not completed_tasks_data and not abandoned_tasks_data and not trash_tasks_data:
return {"error": "无法获取任务数据"}
# 创建Excel文件
excel_buffer = io.BytesIO()
with pd.ExcelWriter(excel_buffer, engine='openpyxl') as writer:
# 处理全部任务
if all_tasks_data:
all_tasks_df = self._process_all_tasks(all_tasks_data)
if not all_tasks_df.empty:
all_tasks_df.to_excel(writer, sheet_name='全部任务', index=False)
app_logger.info(f"全部任务工作表创建完成,共 {len(all_tasks_df)} 条记录")
# 处理已完成任务
if completed_tasks_data:
completed_tasks_df = self._process_completed_tasks(completed_tasks_data)
if not completed_tasks_df.empty:
completed_tasks_df.to_excel(writer, sheet_name='已完成任务', index=False)
app_logger.info(f"已完成任务工作表创建完成,共 {len(completed_tasks_df)} 条记录")
# 处理放弃任务
if abandoned_tasks_data:
abandoned_tasks_df = self._process_abandoned_tasks(abandoned_tasks_data)
if not abandoned_tasks_df.empty:
abandoned_tasks_df.to_excel(writer, sheet_name='放弃任务', index=False)
app_logger.info(f"放弃任务工作表创建完成,共 {len(abandoned_tasks_df)} 条记录")
# 处理垃圾桶任务
if trash_tasks_data:
trash_tasks_df = self._process_trash_tasks(trash_tasks_data)
if not trash_tasks_df.empty:
trash_tasks_df.to_excel(writer, sheet_name='垃圾桶任务', index=False)
app_logger.info(f"垃圾桶任务工作表创建完成,共 {len(trash_tasks_df)} 条记录")
excel_buffer.seek(0)
# 生成文件名
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
filename = f"滴答清单任务导出_{timestamp}.xlsx"
app_logger.info(f"Excel文件生成完成: {filename}")
return {
"filename": filename,
"content_type": "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet",
"content": excel_buffer.getvalue(),
"size": len(excel_buffer.getvalue())
}
except Exception as e:
app_logger.error(f"导出任务到Excel时发生错误: {e}")
return {"error": str(e)}
async def export_focus_records_to_excel(self) -> Dict[str, Any]:
"""
导出专注记录到Excel文件
Returns:
dict: 包含Excel文件内容和元数据的响应
"""
try:
app_logger.info("开始导出专注记录到Excel")
# 获取专注记录数据
focus_timeline_data = await self._get_all_focus_timeline_data()
if not focus_timeline_data:
return {"error": "无法获取专注记录数据"}
# 创建Excel文件
excel_buffer = io.BytesIO()
with pd.ExcelWriter(excel_buffer, engine='openpyxl') as writer:
# 处理专注记录时间线
if focus_timeline_data:
focus_timeline_df = self._process_focus_timeline(focus_timeline_data)
if not focus_timeline_df.empty:
focus_timeline_df.to_excel(writer, sheet_name='专注记录时间线', index=False)
app_logger.info(f"专注记录时间线工作表创建完成,共 {len(focus_timeline_df)} 条记录")
excel_buffer.seek(0)
# 生成文件名
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
filename = f"滴答清单专注记录导出_{timestamp}.xlsx"
app_logger.info(f"专注记录Excel文件生成完成: {filename}")
return {
"filename": filename,
"content_type": "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet",
"content": excel_buffer.getvalue(),
"size": len(excel_buffer.getvalue())
}
except Exception as e:
app_logger.error(f"导出专注记录到Excel时发生错误: {e}")
return {"error": str(e)}
async def _get_all_tasks_data(self) -> Optional[Dict]:
"""获取所有任务数据"""
try:
result = await self.dida_service.get_all_tasks()
if result and 'error' not in result:
return result
return None
except Exception as e:
app_logger.error(f"获取所有任务数据失败: {e}")
return None
async def _get_completed_tasks_data(self) -> Optional[List]:
"""获取已完成任务数据(分页获取所有数据)"""
try:
all_completed_tasks = []
to = None
page_count = 0
while True:
app_logger.info(f"获取已完成任务第 {page_count + 1} 页,to参数: {to}")
result = await self.dida_service.get_completed_tasks(to, "Completed")
if not result or 'error' in result:
app_logger.warning(f"获取已完成任务第 {page_count + 1} 页失败: {result}")
break
if isinstance(result, list) and len(result) > 0:
all_completed_tasks.extend(result)
app_logger.info(f"第 {page_count + 1} 页获取到 {len(result)} 条已完成任务")
# 获取最后一个任务的completedTime作为下次分页参数
last_task = result[-1]
to = last_task.get('completedTime')
if not to:
app_logger.info("最后一个任务没有completedTime,停止分页")
break
page_count += 1
# 如果返回的任务数少于50条,说明已经是最后一页
if len(result) < 50:
app_logger.info("已获取到最后一页已完成任务")
break
else:
app_logger.info("没有更多已完成任务数据")
break
app_logger.info(f"已完成任务分页获取完成,共获取 {len(all_completed_tasks)} 条记录,分 {page_count} 页")
return all_completed_tasks if all_completed_tasks else None
except Exception as e:
app_logger.error(f"获取已完成任务数据失败: {e}")
return None
async def _get_abandoned_tasks_data(self) -> Optional[List]:
"""获取放弃任务数据(分页获取所有数据)"""
try:
all_abandoned_tasks = []
to = None
page_count = 0
while True:
app_logger.info(f"获取放弃任务第 {page_count + 1} 页,to参数: {to}")
result = await self.dida_service.get_completed_tasks(to, "Abandoned")
if not result or 'error' in result:
app_logger.warning(f"获取放弃任务第 {page_count + 1} 页失败: {result}")
break
if isinstance(result, list) and len(result) > 0:
all_abandoned_tasks.extend(result)
app_logger.info(f"第 {page_count + 1} 页获取到 {len(result)} 条放弃任务")
# 获取最后一个任务的completedTime作为下次分页参数
last_task = result[-1]
to = last_task.get('completedTime')
if not to:
app_logger.info("最后一个任务没有completedTime,停止分页")
break
page_count += 1
# 如果返回的任务数少于50条,说明已经是最后一页
if len(result) < 50:
app_logger.info("已获取到最后一页放弃任务")
break
else:
app_logger.info("没有更多放弃任务数据")
break
app_logger.info(f"放弃任务分页获取完成,共获取 {len(all_abandoned_tasks)} 条记录,分 {page_count} 页")
return all_abandoned_tasks if all_abandoned_tasks else None
except Exception as e:
app_logger.error(f"获取放弃任务数据失败: {e}")
return None
async def _get_trash_tasks_data(self) -> Optional[Dict]:
"""获取垃圾桶任务数据"""
try:
result = await self.dida_service.get_trash_tasks()
if result and 'error' not in result:
return result
return None
except Exception as e:
app_logger.error(f"获取垃圾桶任务数据失败: {e}")
return None
async def _get_all_focus_timeline_data(self) -> Optional[List]:
"""获取所有专注记录时间线数据(分页获取所有数据)"""
try:
all_focus_records = []
to_timestamp = None
page_count = 0
max_pages = 100 # 防止无限循环
# 获取认证信息
current_session = self.dida_service.current_session
if not current_session:
app_logger.error("未找到认证会话")
return None
auth_token = current_session['auth_token']
csrf_token = current_session['csrf_token']
while page_count < max_pages:
app_logger.info(f"获取专注记录第 {page_count + 1} 页,to_timestamp: {to_timestamp}")
result = await pomodoro_service.get_focus_timeline(auth_token, csrf_token, to_timestamp)
if not result or 'error' in result:
app_logger.warning(f"获取专注记录第 {page_count + 1} 页失败: {result}")
break
# 检查是否有数据
if not isinstance(result, list) or len(result) == 0:
app_logger.info("没有更多专注记录数据")
break
all_focus_records.extend(result)
app_logger.info(f"第 {page_count + 1} 页获取到 {len(result)} 条专注记录")
# 获取最后一条记录的startTime作为下次分页参数
if len(result) > 0:
last_record = result[-1]
start_time = last_record.get('startTime')
if start_time:
# 转换时间格式用于下次请求
to_timestamp = pomodoro_service._convert_time_to_timestamp(start_time)
else:
app_logger.info("最后一条记录没有startTime,停止分页")
break
else:
break
page_count += 1
# 如果返回的记录数少于31条(通常每页31条),说明已经是最后一页
if len(result) < 31:
app_logger.info("已获取到最后一页专注记录")
break
app_logger.info(f"专注记录分页获取完成,共获取 {len(all_focus_records)} 条记录,分 {page_count} 页")
return all_focus_records if all_focus_records else None
except Exception as e:
app_logger.error(f"获取专注记录时间线数据失败: {e}")
return None
def _process_all_tasks(self, data: Dict) -> pd.DataFrame:
"""处理全部任务数据"""
try:
tasks = data.get('syncTaskBean', {}).get('update', [])
projects = {p['id']: p['name'] for p in data.get('projectProfiles', [])}
processed_tasks = []
for task in tasks:
processed_task = self._flatten_task(task, projects)
processed_tasks.append(processed_task)
return pd.DataFrame(processed_tasks)
except Exception as e:
app_logger.error(f"处理全部任务数据失败: {e}")
return pd.DataFrame()
def _process_completed_tasks(self, data: List) -> pd.DataFrame:
"""处理已完成任务数据"""
try:
processed_tasks = []
for task in data:
processed_task = self._flatten_task(task, {})
processed_tasks.append(processed_task)
return pd.DataFrame(processed_tasks)
except Exception as e:
app_logger.error(f"处理已完成任务数据失败: {e}")
return pd.DataFrame()
def _process_abandoned_tasks(self, data: List) -> pd.DataFrame:
"""处理放弃任务数据"""
try:
processed_tasks = []
for task in data:
processed_task = self._flatten_task(task, {})
processed_tasks.append(processed_task)
return pd.DataFrame(processed_tasks)
except Exception as e:
app_logger.error(f"处理放弃任务数据失败: {e}")
return pd.DataFrame()
def _process_trash_tasks(self, data: Dict) -> pd.DataFrame:
"""处理垃圾桶任务数据"""
try:
tasks = data.get('tasks', [])
processed_tasks = []
for task in tasks:
processed_task = self._flatten_task(task, {})
processed_tasks.append(processed_task)
return pd.DataFrame(processed_tasks)
except Exception as e:
app_logger.error(f"处理垃圾桶任务数据失败: {e}")
return pd.DataFrame()
def _process_focus_timeline(self, data: List) -> pd.DataFrame:
"""处理专注记录时间线数据 - 紧凑型展示"""
try:
processed_records = []
for record in data:
# 为每个专注会话创建一条紧凑记录
compact_record = self._create_compact_focus_record(record)
processed_records.append(compact_record)
return pd.DataFrame(processed_records)
except Exception as e:
app_logger.error(f"处理专注记录时间线数据失败: {e}")
return pd.DataFrame()
def _flatten_task(self, task: Dict, projects: Dict) -> Dict:
"""展平任务数据,包含所有字段"""
try:
flattened = {
# 基本信息
'任务ID': task.get('id', ''),
'任务标题': task.get('title', ''),
'任务内容': task.get('content', ''),
'任务描述': task.get('desc', ''),
'项目ID': task.get('projectId', ''),
'项目名称': projects.get(task.get('projectId', ''), ''),
'排序顺序': task.get('sortOrder', 0),
# 状态和优先级
'任务状态': self._get_status_text(task.get('status', 0)),
'状态代码': task.get('status', 0),
'优先级': task.get('priority', 0),
'完成进度': task.get('progress', 0),
'删除状态': task.get('deleted', 0),
# 时间相关
'创建时间': task.get('createdTime', ''),
'修改时间': task.get('modifiedTime', ''),
'开始日期': task.get('startDate', ''),
'截止日期': task.get('dueDate', ''),
'置顶时间': task.get('pinnedTime', ''),
'完成时间': task.get('completedTime', ''),
'删除时间': task.get('deletedTime', ''),
# 时区和时间设置
'时区': task.get('timeZone', ''),
'是否浮动时间': task.get('isFloating', False),
'是否全天任务': task.get('isAllDay', False),
# 重复设置
'重复任务ID': task.get('repeatTaskId', ''),
'重复标志': task.get('repeatFlag', ''),
'重复来源': task.get('repeatFrom', ''),
'首次重复日期': task.get('repeatFirstDate', ''),
# 提醒设置
'提醒设置': task.get('reminder', ''),
'提醒列表': str(task.get('reminders', [])),
'排除日期': str(task.get('exDate', [])),
# 层级关系
'父任务ID': task.get('parentId', ''),
'子任务ID列表': str(task.get('childIds', [])),
# 其他属性
'标签列表': str(task.get('tags', [])),
'子项目': str(task.get('items', [])),
'附件数量': len(task.get('attachments', [])),
'评论数量': task.get('commentCount', 0),
'列ID': task.get('columnId', ''),
'类型': task.get('kind', ''),
'图片模式': task.get('imgMode', 0),
# 创建者和删除者
'创建者ID': task.get('creator', 0),
'删除者ID': task.get('deletedBy', 0),
# 版本控制
'实体标签': task.get('etag', ''),
# 专注相关
'番茄钟摘要': str(task.get('pomodoroSummaries', [])),
'专注摘要': str(task.get('focusSummaries', [])),
# 附件详情
'附件详情': str(task.get('attachments', [])),
}
return flattened
except Exception as e:
app_logger.error(f"展平任务数据失败: {e}")
return {}
def _get_status_text(self, status_code: int) -> str:
"""获取状态文本描述"""
status_map = {
0: '未完成',
1: '进行中',
2: '已完成',
-1: '已删除'
}
return status_map.get(status_code, f'未知状态({status_code})')
def _create_compact_focus_record(self, record: Dict) -> Dict:
"""创建紧凑型专注记录"""
try:
from datetime import datetime, timedelta
# 基本信息
session_id = record.get('id', '')
session_start = record.get('startTime', '')
session_end = record.get('endTime', '')
pause_duration = record.get('pauseDuration', 0)
# 计算总时长
total_duration = 0
if session_start and session_end:
try:
start_time = datetime.fromisoformat(session_start.replace('Z', '+00:00'))
end_time = datetime.fromisoformat(session_end.replace('Z', '+00:00'))
total_duration = int((end_time - start_time).total_seconds())
except:
total_duration = 0
# 获取任务信息
tasks = record.get('tasks', [])
task_titles = []
project_names = []
for task in tasks:
if task.get('title'):
task_titles.append(task['title'])
if task.get('projectName'):
project_names.append(task['projectName'])
main_task = '; '.join(set(task_titles)) # 去重
main_project = '; '.join(set(project_names)) # 去重
# 生成专注时间段描述
focus_timeline = self._generate_focus_timeline(tasks, pause_duration)
# 生成暂停模式描述
pause_pattern = self._generate_pause_pattern(tasks, pause_duration)
# 格式化会话时间
session_time_str = ""
if session_start and session_end:
try:
start_dt = datetime.fromisoformat(session_start.replace('Z', '+00:00'))
end_dt = datetime.fromisoformat(session_end.replace('Z', '+00:00'))
session_time_str = f"{start_dt.strftime('%Y-%m-%d %H:%M')} - {end_dt.strftime('%H:%M')}"
except:
session_time_str = f"{session_start} - {session_end}"
return {
'会话ID': session_id,
'会话时间': session_time_str,
'总时长': self._format_duration(total_duration),
'暂停时长': self._format_duration(pause_duration),
'任务标题': main_task,
'项目': main_project,
'专注时间段': focus_timeline,
'暂停模式': pause_pattern,
'效率(%)': round((total_duration - pause_duration) / total_duration * 100, 1) if total_duration > 0 else 0,
'时间段数量': len(tasks),
'会话类型': record.get('type', ''),
'实体标签': record.get('etag', '')
}
except Exception as e:
app_logger.error(f"创建紧凑型专注记录失败: {e}")
return {}
def _generate_focus_timeline(self, tasks: List[Dict], total_pause_duration: int) -> str:
"""生成专注时间段描述"""
try:
from datetime import datetime
if not tasks:
return "无专注时间段"
timeline_parts = []
for i, task in enumerate(tasks):
start_time = task.get('startTime', '')
end_time = task.get('endTime', '')
if start_time and end_time:
try:
start_dt = datetime.fromisoformat(start_time.replace('Z', '+00:00'))
end_dt = datetime.fromisoformat(end_time.replace('Z', '+00:00'))
duration = int((end_dt - start_dt).total_seconds())
# 格式化时间段
time_part = f"{start_dt.strftime('%H:%M')}-{end_dt.strftime('%H:%M')}({self._format_duration(duration)})"
timeline_parts.append(time_part)
# 如果不是最后一个时间段,计算暂停时间
if i < len(tasks) - 1:
next_task = tasks[i + 1]
next_start = next_task.get('startTime', '')
if next_start:
try:
next_start_dt = datetime.fromisoformat(next_start.replace('Z', '+00:00'))
pause_duration = int((next_start_dt - end_dt).total_seconds())
if pause_duration > 0:
timeline_parts.append(f"[暂停{self._format_duration(pause_duration)}]")
except:
timeline_parts.append("[暂停未知时长]")
except:
timeline_parts.append(f"时间段{i+1}(解析失败)")
return " → ".join(timeline_parts)
except Exception as e:
app_logger.error(f"生成专注时间段描述失败: {e}")
return "生成失败"
def _generate_pause_pattern(self, tasks: List[Dict], total_pause_duration: int) -> str:
"""生成暂停模式描述"""
try:
if len(tasks) <= 1:
return "无暂停" if total_pause_duration == 0 else f"总暂停{self._format_duration(total_pause_duration)}"
pause_count = len(tasks) - 1
avg_pause = total_pause_duration // pause_count if pause_count > 0 else 0
if pause_count == 1:
return f"暂停1次({self._format_duration(total_pause_duration)})"
else:
return f"暂停{pause_count}次(总计{self._format_duration(total_pause_duration)}, 平均{self._format_duration(avg_pause)})"
except Exception as e:
app_logger.error(f"生成暂停模式描述失败: {e}")
return "分析失败"
def _format_duration(self, seconds: int) -> str:
"""格式化时长显示"""
try:
if seconds < 60:
return f"{seconds}秒"
elif seconds < 3600:
minutes = seconds // 60
remaining_seconds = seconds % 60
if remaining_seconds == 0:
return f"{minutes}分钟"
else:
return f"{minutes}分{remaining_seconds}秒"
else:
hours = seconds // 3600
remaining_minutes = (seconds % 3600) // 60
if remaining_minutes == 0:
return f"{hours}小时"
else:
return f"{hours}小时{remaining_minutes}分钟"
except:
return f"{seconds}秒"
# 创建全局实例
export_service = ExportService()
|
294coder/Efficient-MIF | 1,319 | Pansharpening_Hyper_SR_Matlab_Test_Package/printAllImagesImWriteFR.m | MatrixPrint(:,:,:,1) = I_MS;
MatrixPrint(:,:,:,2) = I_BT_H;
MatrixPrint(:,:,:,3) = I_BDSD;
MatrixPrint(:,:,:,4) = I_C_BDSD;
MatrixPrint(:,:,:,5) = I_BDSD_PC;
MatrixPrint(:,:,:,6) = I_GS;
MatrixPrint(:,:,:,7) = I_GSA;
MatrixPrint(:,:,:,8) = I_C_GSA;
MatrixPrint(:,:,:,9) = I_PRACS;
MatrixPrint(:,:,:,10) = I_AWLP;
MatrixPrint(:,:,:,11) = I_MTF_GLP;
MatrixPrint(:,:,:,12) = I_MTF_GLP_FS;
MatrixPrint(:,:,:,13) = I_MTF_GLP_HPM;
MatrixPrint(:,:,:,14) = I_MTF_GLP_HPM_H;
MatrixPrint(:,:,:,15) = I_MTF_GLP_HPM_R;
MatrixPrint(:,:,:,16) = I_MTF_GLP_CBD;
MatrixPrint(:,:,:,17) = I_C_MTF_GLP_CBD;
MatrixPrint(:,:,:,18) = I_MF;
MatrixPrint(:,:,:,19) = I_FE_HPM;
MatrixPrint(:,:,:,20) = I_SR_D;
MatrixPrint(:,:,:,21) = I_PWMBF;
MatrixPrint(:,:,:,22) = I_TV;
MatrixPrint(:,:,:,23) = I_RR;
MatrixPrint(:,:,:,24) = I_PNN;
MatrixPrint(:,:,:,25) = I_PNN_IDX;
MatrixPrint(:,:,:,26) = I_A_PNN;
MatrixPrint(:,:,:,27) = I_A_PNN_FT;
if size(I_MS,3) == 4
vect_index_RGB = [3,2,1];
else
vect_index_RGB = [5,3,2];
end
titleImages = algorithms;
addpath([pwd,'\Tools']);
figure, MP = showImagesAll(MatrixPrint,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0);
cd 'Outputs'
for ii = 1 : size(MP,4)
imwrite(MP(:,:,:,ii),sprintf('%s.png',algorithms{ii}));
end
imwrite(showPan(I_PAN,0,1,flag_cut_bounds,dim_cut),'PAN.png')
cd .. |
294coder/Efficient-MIF | 2,675 | Pansharpening_Hyper_SR_Matlab_Test_Package/ssim_index.m | function [mssim, ssim_map] = ssim_index(img1, img2, K, window, L)
%========================================================================
if (nargin < 2 || nargin > 5)
mssim = -Inf;
ssim_map = -Inf;
return;
end
if (size(img1) ~= size(img2))
mssim = -Inf;
ssim_map = -Inf;
return;
end
[M N] = size(img1);
if (nargin == 2)
if ((M < 11) || (N < 11))
mssim = -Inf;
ssim_map = -Inf;
return
end
window = fspecial('gaussian', 11, 1.5); %
K(1) = 0.01; % default settings
K(2) = 0.03; %
L = 255; %
end
if (nargin == 3)
if ((M < 11) || (N < 11))
mssim = -Inf;
ssim_map = -Inf;
return
end
window = fspecial('gaussian', 11, 1.5);
L = 255;
if (length(K) == 2)
if (K(1) < 0 || K(2) < 0)
mssim = -Inf;
ssim_map = -Inf;
return;
end
else
mssim = -Inf;
ssim_map = -Inf;
return;
end
end
if (nargin == 4)
[H W] = size(window);
if ((H*W) < 4 || (H > M) || (W > N))
mssim = -Inf;
ssim_map = -Inf;
return
end
L = 255;
if (length(K) == 2)
if (K(1) < 0 || K(2) < 0)
mssim = -Inf;
ssim_map = -Inf;
return;
end
else
mssim = -Inf;
ssim_map = -Inf;
return;
end
end
if (nargin == 5)
[H W] = size(window);
if ((H*W) < 4 || (H > M) || (W > N))
mssim = -Inf;
ssim_map = -Inf;
return
end
if (length(K) == 2)
if (K(1) < 0 || K(2) < 0)
mssim = -Inf;
ssim_map = -Inf;
return;
end
else
mssim = -Inf;
ssim_map = -Inf;
return;
end
end
C1 = (K(1)*L)^2;
C2 = (K(2)*L)^2;
window = window/sum(sum(window));
img1 = double(img1);
img2 = double(img2);
mu1 = filter2(window, img1, 'valid');
mu2 = filter2(window, img2, 'valid');
mu1_sq = mu1.*mu1;
mu2_sq = mu2.*mu2;
mu1_mu2 = mu1.*mu2;
sigma1_sq = filter2(window, img1.*img1, 'valid') - mu1_sq;
sigma2_sq = filter2(window, img2.*img2, 'valid') - mu2_sq;
sigma12 = filter2(window, img1.*img2, 'valid') - mu1_mu2;
if (C1 > 0 && C2 > 0)
ssim_map = ((2*mu1_mu2 + C1).*(2*sigma12 + C2))./((mu1_sq + mu2_sq + C1).*(sigma1_sq + sigma2_sq + C2));
else
numerator1 = 2*mu1_mu2 + C1;
numerator2 = 2*sigma12 + C2;
denominator1 = mu1_sq + mu2_sq + C1;
denominator2 = sigma1_sq + sigma2_sq + C2;
ssim_map = ones(size(mu1));
index = (denominator1.*denominator2 > 0);
ssim_map(index) = (numerator1(index).*numerator2(index))./(denominator1(index).*denominator2(index));
index = (denominator1 ~= 0) & (denominator2 == 0);
ssim_map(index) = numerator1(index)./denominator1(index);
end
mssim = mean2(ssim_map);
return
|
294coder/Efficient-MIF | 1,613 | Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_ref_batched_images.m | function res = analysis_ref_batched_images(path, ratio, full_res, const)
% data should be [0, max_range]
% e.g. wv3: max_range: 2047
data = load(path);
gt = data.gt;
sr = data.sr;
addpath('./Tools')
addpath('./Quality_Indices/')
Q_block_size = 32;
thvalues = 0;
L = 11;
flag_cut_bounds = 0;
if full_res
dim_cut = 21;
else
dim_cut = 30;
end
bs = size(sr, 1);
res = {};
sam = zeros(1, bs);
ergas = zeros(1, bs);
scc = zeros(1, bs);
qn = zeros(1, bs);
qave = zeros(1, bs);
psnr = zeros(1, bs);
ssim = zeros(1, bs);
for i = (1:bs)
sr1 = permute(squeeze(sr(i, :, :, :)), [2, 3, 1]);
gt1 = permute(squeeze(gt(i, :, :, :)), [2, 3, 1]);
[q_avg_gt, sam_gt, ergas_gt, scc_gt, q_gt] = indexes_evaluation(sr1, gt1, ratio, L, Q_block_size, flag_cut_bounds, dim_cut, thvalues);
[psnr_gt, ssim_gt] = quality_assess(sr1 / const, gt1 / const);
sam(i) = sam_gt;
ergas(i) = ergas_gt;
scc(i) = scc_gt;
qn(i) = q_gt;
qave(i) = q_avg_gt;
psnr(i) = psnr_gt;
ssim(i) = ssim_gt;
fprintf("sample %d - sam: %f, ergas: %f, scc: %f, qn: %f, q_ave: %f, psnr: %f, ssim: %f \n", i, sam_gt, ergas_gt, scc_gt, q_gt, q_avg_gt, psnr_gt, ssim_gt)
end
res.sam = [mean(sam), std(sam)];
res.ergas = [mean(ergas), std(ergas)];
res.scc = [mean(scc), std(scc)];
res.qn = [mean(qn), std(qn)];
res.qave = [mean(qave), std(qave)];
res.psnr = [mean(psnr), std(psnr)];
res.ssim = [mean(ssim), std(ssim)];
end
|
281677160/openwrt-package | 6,592 | luci-app-passwall/root/usr/share/passwall/0_default_config |
config global
option enabled '0'
option socks_enabled '0'
option tcp_node_socks_port '1070'
option filter_proxy_ipv6 '1'
option dns_shunt 'chinadns-ng'
option dns_mode 'tcp'
option remote_dns '1.1.1.1'
list smartdns_remote_dns 'https://1.1.1.1/dns-query'
option use_default_dns 'direct'
option chinadns_ng_default_tag 'none'
option dns_redirect '1'
option use_direct_list '1'
option use_proxy_list '1'
option use_block_list '1'
option use_gfw_list '1'
option chn_list 'direct'
option tcp_proxy_mode 'proxy'
option udp_proxy_mode 'proxy'
option localhost_proxy '1'
option client_proxy '1'
option acl_enable '0'
option log_tcp '0'
option log_udp '0'
option loglevel 'error'
option trojan_loglevel '4'
option log_chinadns_ng '0'
config global_haproxy
option balancing_enable '0'
config global_delay
option start_daemon '1'
option start_delay '60'
config global_forwarding
option tcp_no_redir_ports 'disable'
option udp_no_redir_ports 'disable'
option tcp_proxy_drop_ports 'disable'
option udp_proxy_drop_ports '443'
option tcp_redir_ports '22,25,53,80,143,443,465,587,853,873,993,995,5222,8080,8443,9418'
option udp_redir_ports '1:65535'
option accept_icmp '0'
option use_nft '0'
option tcp_proxy_way 'redirect'
option ipv6_tproxy '0'
config global_xray
option sniffing_override_dest '0'
config global_singbox
option sniff_override_destination '0'
config global_other
option auto_detection_time 'tcping'
option show_node_info '0'
config global_rules
option auto_update '0'
option chnlist_update '1'
option chnroute_update '1'
option chnroute6_update '1'
option gfwlist_update '1'
option geosite_update '0'
option geoip_update '0'
list gfwlist_url 'https://fastly.jsdelivr.net/gh/Loyalsoldier/v2ray-rules-dat@release/gfw.txt'
list chnroute_url 'https://ispip.clang.cn/all_cn.txt'
list chnroute_url 'https://fastly.jsdelivr.net/gh/gaoyifan/china-operator-ip@ip-lists/china.txt'
list chnroute6_url 'https://ispip.clang.cn/all_cn_ipv6.txt'
list chnroute6_url 'https://fastly.jsdelivr.net/gh/gaoyifan/china-operator-ip@ip-lists/china6.txt'
list chnlist_url 'https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/accelerated-domains.china.conf'
list chnlist_url 'https://fastly.jsdelivr.net/gh/felixonmars/dnsmasq-china-list/apple.china.conf'
option v2ray_location_asset '/usr/share/v2ray/'
option geoip_url 'https://github.com/Loyalsoldier/geoip/releases/latest/download/geoip.dat'
option geosite_url 'https://github.com/Loyalsoldier/v2ray-rules-dat/releases/latest/download/geosite.dat'
config global_app
option sing_box_file '/usr/bin/sing-box'
option xray_file '/usr/bin/xray'
option hysteria_file '/usr/bin/hysteria'
config global_subscribe
option filter_keyword_mode '1'
list filter_discard_list '距离下次重置剩余'
list filter_discard_list '套餐到期'
list filter_discard_list '过期时间'
list filter_discard_list '剩余流量'
list filter_discard_list 'QQ群'
list filter_discard_list '官网'
config nodes 'myshunt'
option remarks '分流总节点'
option type 'Xray'
option protocol '_shunt'
option DirectGame '_direct'
option ProxyGame '_default'
option AIGC '_default'
option Streaming '_default'
option Proxy '_default'
option Direct '_direct'
option default_node '_direct'
option domainStrategy 'IPOnDemand'
config shunt_rules 'DirectGame'
option remarks 'DirectGame'
option domain_list '# steam直连域名获取国内CDN走国内线路下载
cm.steampowered.com
steamserver.net
# steam国内CDN华为云
steampipe.steamcontent.tnkjmec.com
# steam国内CDN白山云
st.dl.eccdnx.com
st.dl.bscstorage.net
st.dl.pinyuncloud.com
# steam国内CDN新流云(原金山云)(支持ipv6)
dl.steam.clngaa.com
# steam国内CDN网宿
cdn.mileweb.cs.steampowered.com.8686c.com
cdn-ws.content.steamchina.com
# steam国内CDN腾讯云 (蒸汽中国独占)
cdn-qc.content.steamchina.com
# steam国内CDN阿里云(支持ipv6)
cdn-ali.content.steamchina.com
xz.pphimalayanrt.com
lv.queniujq.cn
alibaba.cdn.steampipe.steamcontent.com
# 国内游戏geosite域名
geosite:category-games@cn'
option ip_list '# steam直连IP
45.121.184.0/24
103.10.124.0/23
103.28.54.0/24
146.66.152.0/24
146.66.155.0/24
153.254.86.0/24
155.133.224.0/22
155.133.230.0/24
155.133.232.0/23
155.133.234.0/24
155.133.236.0/22
155.133.240.0/23
155.133.244.0/23
155.133.246.0/24
155.133.248.0/21
162.254.192.0/21
185.25.182.0/23
190.217.32.0/22
192.69.96.0/22
205.196.6.0/24
208.64.200.0/22
208.78.164.0/22
205.185.194.0/24'
config shunt_rules 'ProxyGame'
option remarks 'ProxyGame'
option domain_list '# steam 商店/客服/聊天/网页布局/API/二维码/Google云同步 代理URL
steamcommunity.com
www.steamcommunity.com
store.steampowered.com
checkout.steampowered.com
api.steampowered.com
help.steampowered.com
login.steampowered.com
store.akamai.steamstatic.com
steambroadcast.akamaized.net
steamvideo-a.akamaihd.net
steamusercontent-a.akamaihd.net
steamstore-a.akamaihd.net
steamcommunity-a.akamaihd.net
steamcdn-a.akamaihd.net
steamuserimages-a.akamaihd.net
community.akamai.steamstatic.com
avatars.akamai.steamstatic.com
community.steamstatic.com
cdn.akamai.steamstatic.com
avatars.steamstatic.com
shared.akamai.steamstatic.com
clan.akamai.steamstatic.com
cdn.cloudflare.steamstatic.com
community.cloudflare.steamstatic.com
store.cloudflare.steamstatic.com
avatars.cloudflare.steamstatic.com
clan.cloudflare.steamstatic.com
shared.cloudflare.steamstatic.com
steam-chat.com
steamcloud-ugc.storage.googleapis.com
steamcloud-eu-ams.storage.googleapis.com
steamcloud-eu-fra.storage.googleapis.com
steamcloud-finland.storage.googleapis.com
steamcloud-saopaulo.storage.googleapis.com
steamcloud-singapore.storage.googleapis.com
steamcloud-sydney.storage.googleapis.com
steamcloud-taiwan.storage.googleapis.com
steamcloud-eu.storage.googleapis.com
geosite:category-games'
config shunt_rules 'AIGC'
option remarks 'AIGC'
option domain_list 'geosite:category-ai-!cn
geosite:apple-intelligence'
config shunt_rules 'Streaming'
option remarks 'Streaming'
option domain_list 'geosite:netflix
geosite:disney'
config shunt_rules 'Proxy'
option remarks 'Proxy'
option domain_list 'geosite:geolocation-!cn'
option ip_list '149.154.160.0/20
91.108.4.0/22
91.108.56.0/24
109.239.140.0/24
67.198.55.0/24
8.8.4.4
8.8.8.8
208.67.222.222
208.67.220.220
1.1.1.1
1.1.1.2
1.0.0.1
9.9.9.9
149.112.112.112
2001:67c:4e8::/48
2001:b28:f23c::/48
2001:b28:f23d::/48
2001:b28:f23f::/48
2001:b28:f242::/48
2001:4860:4860::8888
2001:4860:4860::8844
2606:4700:4700::1111
2606:4700:4700::1001'
config shunt_rules 'Direct'
option remarks 'Direct'
option domain_list 'geosite:cn'
option ip_list '223.5.5.5/32
223.6.6.6/32
119.29.29.29/32
180.76.76.76/32
114.114.114.114/32
114.114.115.115/32
1.12.12.12/32
120.53.53.53/32
geoip:cn
geoip:private'
|
2977094657/DidaAPI | 16,185 | services/wechat_service.py | """微信登录服务模块"""
import re
import uuid
from typing import Optional, Dict, Any, Tuple
import httpx
from utils import app_logger
from core import config, db, urls
from models import WeChatQRResponse, WeChatValidateResponse, PasswordLoginRequest
class WeChatLoginService:
"""微信登录服务类"""
def __init__(self):
self.request_config = config.get('request_config', {})
self.client = httpx.AsyncClient(timeout=self.request_config.get('timeout', 30.0))
async def get_qr_code(self, state: str = "Lw==") -> Optional[WeChatQRResponse]:
"""
获取微信登录二维码
Args:
state: 状态参数,默认为 "Lw=="
Returns:
WeChatQRResponse: 包含二维码URL和密钥的响应对象
"""
try:
# 使用统一的URL构建函数
qr_url = urls.build_wechat_qr_url(state)
app_logger.info(f"请求微信二维码: {qr_url}")
# 发送请求
response = await self.client.get(qr_url)
response.raise_for_status()
# 记录完整响应
app_logger.debug(f"微信二维码响应状态: {response.status_code}")
app_logger.debug(f"微信二维码响应头: {dict(response.headers)}")
app_logger.debug(f"微信二维码响应内容长度: {len(response.text)}")
# 解析HTML中的二维码图片链接
qr_code_key = self._extract_qr_code_key(response.text)
if not qr_code_key:
app_logger.error("未能从响应中提取二维码密钥")
return None
# 构建完整的二维码图片URL
qr_code_url = f"{urls.WECHAT_URLS['qr_image_base_url']}/{qr_code_key}"
# 记录到数据库
db.log_wechat_login(qr_code_key=qr_code_key, state=state)
app_logger.info(f"成功获取二维码: {qr_code_url}")
return WeChatQRResponse(
qr_code_url=qr_code_url,
qr_code_key=qr_code_key,
state=state
)
except Exception as e:
app_logger.error(f"获取微信二维码失败: {e}")
return None
def _extract_qr_code_key(self, html_content: str) -> Optional[str]:
"""
从HTML内容中提取二维码密钥
Args:
html_content: HTML响应内容
Returns:
str: 16位二维码密钥,如果未找到则返回None
"""
try:
# 查找二维码图片标签
pattern = r'<img[^>]*class="[^"]*qrcode[^"]*"[^>]*src="([^"]*)"'
match = re.search(pattern, html_content)
if match:
src_url = match.group(1)
app_logger.debug(f"找到二维码图片src: {src_url}")
# 提取最后16位字符
qr_code_key = src_url.split('/')[-1]
if len(qr_code_key) >= 16:
qr_code_key = qr_code_key[-16:]
app_logger.info(f"提取到二维码密钥: {qr_code_key}")
return qr_code_key
else:
app_logger.warning(f"二维码密钥长度不足16位: {qr_code_key}")
# 如果上面的方法失败,尝试其他模式
patterns = [
r'/connect/qrcode/([a-zA-Z0-9]{16})',
r'qrcode/([a-zA-Z0-9]{16})',
r'src="[^"]*?([a-zA-Z0-9]{16})"'
]
for pattern in patterns:
matches = re.findall(pattern, html_content)
if matches:
qr_code_key = matches[0]
app_logger.info(f"通过备用模式提取到二维码密钥: {qr_code_key}")
return qr_code_key
app_logger.error("未能从HTML中提取二维码密钥")
app_logger.debug(f"HTML内容片段: {html_content[:500]}...")
return None
except Exception as e:
app_logger.error(f"提取二维码密钥时发生错误: {e}")
return None
async def poll_qr_status(self, qr_code_key: str, max_attempts: int = 60) -> Optional[WeChatValidateResponse]:
"""
轮询二维码状态,检查是否已扫码登录
Args:
qr_code_key: 二维码密钥
max_attempts: 最大轮询次数,默认60次(约5分钟)
Returns:
WeChatValidateResponse: 登录结果
"""
import asyncio
for attempt in range(max_attempts):
try:
app_logger.info(f"轮询二维码状态,第 {attempt + 1}/{max_attempts} 次")
# 使用统一的URL构建函数
poll_url = urls.build_wechat_poll_url(qr_code_key)
headers = {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36',
'Referer': 'https://open.weixin.qq.com/',
'Accept': '*/*',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8'
}
response = await self.client.get(poll_url, headers=headers)
response_text = response.text
app_logger.debug(f"轮询响应: {response_text}")
# 解析微信轮询响应
# 实际响应格式是: window.wx_errcode=405;window.wx_code='xxx';
if "window.wx_errcode" in response_text:
import re
# 提取错误码
errcode_match = re.search(r'window\.wx_errcode\s*=\s*(\d+)', response_text)
code_match = re.search(r"window\.wx_code\s*=\s*'([^']*)'", response_text)
if errcode_match:
errcode = int(errcode_match.group(1))
wx_code = code_match.group(1) if code_match else ''
app_logger.info(f"轮询状态 - errcode: {errcode}, wx_code: {wx_code}")
if errcode == 405 and wx_code:
# 登录成功,获得了授权码
app_logger.info(f"检测到登录成功,获得授权码: {wx_code}")
# 使用获得的code进行验证
return await self.validate_wechat_login(wx_code)
elif errcode == 404:
app_logger.info("等待扫码...")
elif errcode == 403:
app_logger.info("二维码已扫描,等待用户确认")
elif errcode == 408:
app_logger.info("二维码已过期")
break
elif errcode == 400:
app_logger.info("二维码已失效")
break
# 等待5秒后继续轮询
await asyncio.sleep(5)
except Exception as e:
app_logger.error(f"轮询二维码状态失败: {e}")
await asyncio.sleep(5)
app_logger.warning("轮询超时,未检测到登录")
return WeChatValidateResponse(
success=False,
message="轮询超时,请重新获取二维码",
token=None,
user_info=None,
cookies=None,
raw_response={"error": "polling_timeout"}
)
async def validate_wechat_login(self, code: str, state: str = "Lw==") -> Optional[WeChatValidateResponse]:
"""
验证微信登录
Args:
code: 扫码后获得的验证码
state: 状态参数
Returns:
WeChatValidateResponse: 验证响应对象
"""
try:
# 使用统一的URL构建函数
validate_url = urls.build_wechat_validate_url(code, state)
app_logger.info(f"验证微信登录: {validate_url}")
# 设置请求头,模拟浏览器请求
headers = {
'Accept': '*/*',
'Accept-Encoding': 'gzip, deflate, br, zstd',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7',
'Content-Type': 'application/json',
'Origin': 'https://dida365.com',
'Referer': 'https://dida365.com/',
'Sec-Ch-Ua': '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
'Sec-Ch-Ua-Mobile': '?0',
'Sec-Ch-Ua-Platform': '"Windows"',
'Sec-Fetch-Dest': 'empty',
'Sec-Fetch-Mode': 'cors',
'Sec-Fetch-Site': 'same-site',
'User-Agent': self.request_config.get('user_agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'),
'X-Device': self.request_config.get('device_info', '{}')
}
# 发送验证请求
response = await self.client.get(validate_url, headers=headers)
# 记录详细的响应信息
app_logger.info(f"验证响应状态码: {response.status_code}")
app_logger.info(f"验证响应头: {dict(response.headers)}")
# 提取cookies
cookies = {}
if hasattr(response, 'cookies') and response.cookies:
for cookie_name, cookie_value in response.cookies.items():
cookies[cookie_name] = cookie_value
# 同时从Set-Cookie头中解析cookies
set_cookie_header = response.headers.get('set-cookie', '')
if set_cookie_header:
import re
# 解析Set-Cookie头
cookie_matches = re.findall(r'([^=]+)=([^;]+)', set_cookie_header)
for name, value in cookie_matches:
cookies[name.strip()] = value.strip()
app_logger.info(f"响应cookies: {cookies}")
# 尝试解析JSON响应
response_data = {}
try:
response_data = response.json()
app_logger.info(f"验证响应JSON: {response_data}")
except Exception as json_error:
app_logger.warning(f"响应不是有效的JSON: {json_error}")
app_logger.info(f"响应文本内容: {response.text}")
response_data = {"raw_text": response.text}
# 检查是否成功
success = response.status_code == 200
# 提取认证令牌
token = cookies.get('t', '')
csrf_token = cookies.get('_csrf_token', '')
# 保存会话信息
if success and token:
session_id = str(uuid.uuid4())
session_data = {
'session_id': session_id,
'token': token,
'csrf_token': csrf_token,
'cookies': cookies,
'is_active': True
}
db.save_user_session(session_data)
# 自动设置滴答清单API认证会话
try:
from services.dida_service import dida_service
dida_service.set_auth_session(token, csrf_token)
app_logger.info("已自动设置滴答清单API认证会话")
except Exception as e:
app_logger.warning(f"自动设置滴答清单API认证会话失败: {e}")
# 记录登录日志
db.log_wechat_login(
qr_code_key="", # 这里可能需要从之前的记录中关联
validation_code=code,
state=state,
response_data={
'status_code': response.status_code,
'headers': dict(response.headers),
'cookies': cookies,
'json_data': response_data
},
status='success' if success else 'failed'
)
return WeChatValidateResponse(
success=success,
message="登录成功" if success else "登录失败",
token=token if token else None,
user_info=response_data.get('user', {}),
cookies=cookies,
raw_response=response_data
)
except Exception as e:
app_logger.error(f"验证微信登录失败: {e}")
# 记录失败日志
db.log_wechat_login(
qr_code_key="",
validation_code=code,
state=state,
response_data={'error': str(e)},
status='failed'
)
return WeChatValidateResponse(
success=False,
message=f"验证失败: {str(e)}",
token=None,
user_info=None,
cookies=None,
raw_response={'error': str(e)}
)
async def password_login(self, username: str, password: str) -> dict:
"""
密码登录滴答清单
Args:
username: 登录账户(邮箱或手机号)
password: 登录密码
Returns:
dict: 原始响应数据
"""
try:
# 使用统一的URL构建函数
login_url = urls.build_password_login_url(wc=True, remember=True)
app_logger.info(f"密码登录请求: {login_url}")
# 构建请求体
login_data = {
"username": username,
"password": password
}
# 设置请求头,模拟浏览器请求
headers = {
'Accept': 'application/json, text/plain, */*',
'Accept-Encoding': 'gzip, deflate, br, zstd',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7',
'Content-Type': 'application/json',
'Origin': 'https://dida365.com',
'Referer': 'https://dida365.com/',
'Sec-Ch-Ua': '"Chromium";v="136", "Google Chrome";v="136", "Not.A/Brand";v="99"',
'Sec-Ch-Ua-Mobile': '?0',
'Sec-Ch-Ua-Platform': '"Windows"',
'Sec-Fetch-Dest': 'empty',
'Sec-Fetch-Mode': 'cors',
'Sec-Fetch-Site': 'same-site',
'User-Agent': self.request_config.get('user_agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'),
'X-Device': self.request_config.get('device_info', '{}')
}
# 发送POST请求
response = await self.client.post(login_url, json=login_data, headers=headers)
# 记录详细的响应信息
app_logger.info(f"密码登录响应状态码: {response.status_code}")
app_logger.info(f"密码登录响应头: {dict(response.headers)}")
# 提取cookies
cookies = {}
if hasattr(response, 'cookies') and response.cookies:
for cookie_name, cookie_value in response.cookies.items():
cookies[cookie_name] = cookie_value
# 同时从Set-Cookie头中解析cookies
set_cookie_header = response.headers.get('set-cookie', '')
if set_cookie_header:
import re
# 解析Set-Cookie头
cookie_matches = re.findall(r'([^=]+)=([^;]+)', set_cookie_header)
for name, value in cookie_matches:
cookies[name.strip()] = value.strip()
app_logger.info(f"密码登录响应cookies: {cookies}")
# 尝试解析JSON响应
response_data = {}
try:
response_data = response.json()
app_logger.info(f"密码登录响应JSON: {response_data}")
except Exception as json_error:
app_logger.warning(f"响应不是有效的JSON: {json_error}")
app_logger.info(f"响应文本内容: {response.text}")
response_data = {"raw_text": response.text}
# 检查是否成功
success = response.status_code == 200 and 'token' in response_data
# 如果成功,保存会话信息
if success:
token = response_data.get('token', '')
if token:
session_id = str(uuid.uuid4())
session_data = {
'session_id': session_id,
'token': token,
'csrf_token': '', # 密码登录可能不返回CSRF token
'cookies': cookies,
'is_active': True
}
db.save_user_session(session_data)
# 自动设置滴答清单API认证会话
try:
from services.dida_service import dida_service
dida_service.set_auth_session(token, '')
app_logger.info("已自动设置滴答清单API认证会话")
except Exception as e:
app_logger.warning(f"自动设置滴答清单API认证会话失败: {e}")
# 直接返回原始响应
return response_data
except Exception as e:
app_logger.error(f"密码登录失败: {e}")
# 返回错误响应
return {'error': str(e)}
async def close(self):
"""关闭HTTP客户端"""
await self.client.aclose()
# 全局微信登录服务实例
wechat_service = WeChatLoginService()
|
2977094657/DidaAPI | 5,269 | services/habit_service.py | """习惯管理服务模块"""
import httpx
from typing import Optional
from utils import app_logger
from core import urls
# 不再使用响应模型,直接返回原始响应
class HabitService:
"""习惯管理服务类"""
def __init__(self):
self.client = httpx.AsyncClient(timeout=30.0)
def _build_auth_headers(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证请求头"""
return {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36',
'Accept': 'application/json, text/plain, */*',
'Accept-Language': 'zh-CN,zh;q=0.9,en;q=0.8',
'Content-Type': 'application/json',
'X-Requested-With': 'XMLHttpRequest',
'X-Tz': 'Asia/Shanghai',
}
def _build_auth_cookies(self, auth_token: str, csrf_token: str) -> dict:
"""构建认证cookies"""
return {
't': auth_token,
'_csrf_token': csrf_token
}
async def get_habits(self, auth_token: str, csrf_token: str) -> dict:
"""
获取习惯列表
Args:
auth_token: 认证令牌
csrf_token: CSRF令牌
Returns:
dict: 原始响应数据
"""
try:
url = urls.build_dida_api_url(urls.DIDA_HABIT_APIS["get_habits"])
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
app_logger.info(f"请求获取习惯列表: {url}")
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
response_data = response.json()
app_logger.info(f"成功获取习惯列表,习惯数: {len(response_data) if isinstance(response_data, list) else 0}")
# 直接返回原始响应
return response_data
else:
app_logger.error(f"获取习惯列表失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"获取习惯列表时发生错误: {e}")
return {"error": str(e)}
async def get_week_current_statistics(self, auth_token: str, csrf_token: str) -> dict:
"""
获取本周习惯打卡统计
Args:
auth_token: 认证令牌
csrf_token: CSRF令牌
Returns:
dict: 原始响应数据
"""
try:
url = urls.build_dida_api_url(urls.DIDA_HABIT_APIS["week_current_statistics"])
headers = self._build_auth_headers(auth_token, csrf_token)
cookies = self._build_auth_cookies(auth_token, csrf_token)
app_logger.info(f"请求获取本周习惯打卡统计: {url}")
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
response_data = response.json()
app_logger.info("成功获取本周习惯打卡统计")
# 直接返回原始响应
return response_data
else:
app_logger.error(f"获取本周习惯打卡统计失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"获取本周习惯打卡统计时发生错误: {e}")
return {"error": str(e)}
async def export_habits(self, auth_token: str, csrf_token: str) -> dict:
"""
导出习惯数据(Excel格式)
Args:
auth_token: 认证令牌
csrf_token: CSRF令牌
Returns:
dict: 包含文件内容和元数据的响应
"""
try:
url = urls.build_dida_api_url(urls.DIDA_HABIT_APIS["export_habits"])
# 对于文件下载,需要修改Accept头
headers = self._build_auth_headers(auth_token, csrf_token)
headers['Accept'] = '*/*'
headers['X-CSRFToken'] = csrf_token
cookies = self._build_auth_cookies(auth_token, csrf_token)
app_logger.info(f"请求导出习惯数据: {url}")
response = await self.client.get(url, headers=headers, cookies=cookies)
if response.status_code == 200:
# 获取文件名
content_disposition = response.headers.get('content-disposition', '')
filename = 'habits_export.xlsx'
if 'filename=' in content_disposition:
filename = content_disposition.split('filename=')[1].split(';')[0].strip('"')
app_logger.info(f"成功导出习惯数据,文件名: {filename}")
# 返回文件内容和元数据
return {
"filename": filename,
"content_type": response.headers.get('content-type', 'application/vnd.openxmlformats-officedocument.spreadsheetml.sheet'),
"content": response.content,
"size": len(response.content)
}
else:
app_logger.error(f"导出习惯数据失败,状态码: {response.status_code}")
return {"error": f"HTTP {response.status_code}", "text": response.text}
except Exception as e:
app_logger.error(f"导出习惯数据时发生错误: {e}")
return {"error": str(e)}
async def close(self):
"""关闭HTTP客户端"""
await self.client.aclose()
# 全局习惯服务实例
habit_service = HabitService()
|
294coder/Efficient-MIF | 7,005 | Pansharpening_Hyper_SR_Matlab_Test_Package/rantest.m | %This is a demo to run fusion algorithms on Reduced Resolution
% LJ Deng(UESTC) TJ Zhang
% 2022-05-07
clear; close all;
%% =======load directors========
% Tools
addpath([pwd,'/Tools']);
% Select algorithms to run
algorithms = {'GT','EXP','BT-H','BDSD-PC','C-GSA','SR-D',...
'MTF-GLP-HPM-R','MTF-GLP-FS','TV','PanNet','PNN','DiCNN','FusionNet','LAGConv','MSDCNN','BDPN'};%'PNN'
location1 = [2 40 4 43]; %default: data6: [10 50 1 60]; data7:[140 180 5 60]
location2 = []; %default: data6: [190 240 5 60]; data7:[190 235 120 150]
sensor = 'WV3';
%% =======read Multiple TestData_wv3.h5 (four 512x512 WV3 simulated data)========
file_test = '/Data2/DataSet/pansharpening_2/test_data/WV3/test_wv3_multiExm1.h5';
gt_multiExm_tmp = h5read(file_test,'/gt'); % WxHxCxN=1x2x3x4
gt_multiExm = permute(gt_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
ms_multiExm_tmp = h5read(file_test,'/ms'); % WxHxCxN=1x2x3x4
ms_multiExm = permute(ms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
lms_multiExm_tmp = h5read(file_test,'/lms'); % WxHxCxN=1x2x3x4
lms_multiExm = permute(lms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
pan_multiExm_tmp = h5read(file_test,'/pan'); % WxHxCxN=1x2x3x4
pan_multiExm = permute(pan_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
data_name = '3_EPS/WV3/multi/'; % director to save EPS figures
%% ==========Read each Data====================
exm_num = size(ms_multiExm, 1);
for i = 1 : 20% i = 1 or 2 ...
%% read each data
HRMS_tmp = gt_multiExm(i, :, :, :); % I_GT
I_GT = squeeze(HRMS_tmp);
LRMS_tmp = ms_multiExm(i, :, :, :); % I_MS_LR
I_MS_LR = squeeze(LRMS_tmp);
LMS_tmp = lms_multiExm(i, :, :, :); % I_MS
I_MS = squeeze(LMS_tmp);
PAN_tmp = pan_multiExm(i, :, :, :); % I_PAN
I_PAN = squeeze(PAN_tmp);
%% Initialization of the Matrix of Results
NumIndexes = 5;
MatrixResults = zeros(numel(algorithms),NumIndexes);
alg = 0;
%% load Indexes for WV3_RR
sensor = 'WV3';
Qblocks_size = 32;
bicubic = 0;% Interpolator
flag_cut_bounds = 0;% Cut Final Image
dim_cut = 21;% Cut Final Image
thvalues = 0;% Threshold values out of dynamic range
printEPS = 0;% Print Eps
ratio = 4;% Resize Factor
L = 11;% Radiometric Resolution
%% show I_MS_LR, I_GT, PAN Imgs:
showImage8_zoomin(I_MS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
showPan(I_PAN,printEPS,2,flag_cut_bounds,dim_cut);
print('-deps', strcat(data_name, num2str(i-1), '_pan', '.eps'))
%% ======GT ===================
if ismember('GT',algorithms)
alg = alg + 1;
[Q_avg_GT, SAM_GT, ERGAS_GT, SCC_GT, Q_GT] = indexes_evaluation(I_GT,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_GT,Q_avg_GT,SAM_GT,ERGAS_GT,SCC_GT];
MatrixImage(:,:,:,alg) = I_GT;
Q_avg_GT_multiexm(i) = Q_avg_GT;
SAM_GT_multiexm(i) = SAM_GT;
ERGAS_GT_multiexm(i) = ERGAS_GT;
SCC_GT_multiexm(i) = SCC_GT;
Q_GT_multiexm(i) = Q_GT;
showImage8_zoomin(I_GT,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
print('-depsc', strcat(data_name, num2str(i-1), '_gt', '.eps'))
end
%% ====== 4) BDPN Method ======
file_bdpn = 'bdpn_wv3_rs';
load(strcat('/Data2/DataSet/pansharpening_2/results/p2/wv3_multiExm1.h5/BDPN/Test/model_2022-05-21-10-46-03/results/output_mulExm_' , num2str(i-1), '.mat')) % load i-th image for DiCNN
I_bdpn = double(sr);
if ismember('BDPN',algorithms)
alg = alg + 1;
[Q_avg_bdpn, SAM_bdpn, ERGAS_bdpn, SCC_bdpn, Q_bdpn] = indexes_evaluation(I_bdpn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_bdpn,Q_avg_bdpn,SAM_bdpn,ERGAS_bdpn,SCC_bdpn];
MatrixImage(:,:,:,alg) = I_bdpn;
Q_avg_bdpn_multiexm(i) = Q_avg_bdpn;
SAM_bdpn_multiexm(i) = SAM_bdpn;
ERGAS_bdpn_multiexm(i) = ERGAS_bdpn;
SCC_bdpn_multiexm(i) = SCC_bdpn;
Q_bdpn_multiexm(i) = Q_bdpn;
showImage8_zoomin(I_bdpn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
print('-depsc', strcat(data_name, num2str(i-1),'_bdpn.eps'))
end
%
%
% %% ====== 7) APNN Method ======
% file_apnn = 'apnn_wv3_rs';
% load(strcat('2_DL_Result/WV3/APNN/', file_apnn, num2str(i-1), '.mat')) % load i-th image for DiCNN
% I_apnn = 2047*double(apnn_wv3_rs);
%
% if ismember('APNN',algorithms)
% alg = alg + 1;
% [Q_avg_apnn, SAM_apnn, ERGAS_apnn, SCC_apnn, Q_apnn] = indexes_evaluation(I_apnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_apnn,Q_avg_apnn,SAM_apnn,ERGAS_apnn,SCC_apnn];
% MatrixImage(:,:,:,alg) = I_apnn;
%
% Q_avg_apnn_multiexm(i) = Q_avg_apnn;
% SAM_apnn_multiexm(i) = SAM_apnn;
% ERGAS_apnn_multiexm(i) = ERGAS_apnn;
% SCC_apnn_multiexm(i) = SCC_apnn;
% Q_apnn_multiexm(i) = Q_apnn;
%
% showImage8_zoomin(I_apnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% print('-depsc', strcat(data_name, num2str(i-1),'_apnn.eps'))
% end
end
%% Print in LATEX
%% View All
if size(I_GT,3) == 4
vect_index_RGB = [3,2,1];
else
vect_index_RGB = [5,3,2];
end
titleImages = algorithms;
figure, showImagesAll(MatrixImage,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0);
%% ======Display the final average performance =======
% GT: average Q_avg
% bdpn: average Q_avg
avg_Q_bdpn_multiexm = mean(Q_bdpn_multiexm);
std_Q_bdpn_multiexm = std(Q_bdpn_multiexm);
avg_Q_avg_bdpn_multiexm = mean(Q_avg_bdpn_multiexm);
std_Q_avg_bdpn_multiexm = std(Q_avg_bdpn_multiexm);
avg_SAM_bdpn_multiexm = mean(SAM_bdpn_multiexm);
std_SAM_bdpn_multiexm = std(SAM_bdpn_multiexm);
avg_ERGAS_bdpn_multiexm = mean(ERGAS_bdpn_multiexm);
std_ERGAS_bdpn_multiexm = std(ERGAS_bdpn_multiexm);
avg_SCC_bdpn_multiexm = mean(SCC_bdpn_multiexm);
std_SCC_bdpn_multiexm = std(SCC_bdpn_multiexm);
Avg_MatrixResults(17,:) = [avg_Q_bdpn_multiexm, std_Q_bdpn_multiexm, avg_Q_avg_bdpn_multiexm, std_Q_avg_bdpn_multiexm, ...
avg_SAM_bdpn_multiexm, std_SAM_bdpn_multiexm, avg_ERGAS_bdpn_multiexm, std_ERGAS_bdpn_multiexm,...
avg_SCC_bdpn_multiexm, std_SCC_bdpn_multiexm];
fprintf('\n')
disp('#######################################################')
disp(['Display the performance for:', num2str(1:i)])
disp('#######################################################')
disp(' |====Q====|===Q_avg===|=====SAM=====|======ERGAS=======|=======SCC=======')
MatrixResults |
294coder/Efficient-MIF | 1,349 | Pansharpening_Hyper_SR_Matlab_Test_Package/printAllImagesImWriteRR.m | MatrixPrint(:,:,:,1) = I_GT;
MatrixPrint(:,:,:,2) = I_MS;
MatrixPrint(:,:,:,3) = I_BT_H;
MatrixPrint(:,:,:,4) = I_BDSD;
MatrixPrint(:,:,:,5) = I_C_BDSD;
MatrixPrint(:,:,:,6) = I_BDSD_PC;
MatrixPrint(:,:,:,7) = I_GS;
MatrixPrint(:,:,:,8) = I_GSA;
MatrixPrint(:,:,:,9) = I_C_GSA;
MatrixPrint(:,:,:,10) = I_PRACS;
MatrixPrint(:,:,:,11) = I_AWLP;
MatrixPrint(:,:,:,12) = I_MTF_GLP;
MatrixPrint(:,:,:,13) = I_MTF_GLP_FS;
MatrixPrint(:,:,:,14) = I_MTF_GLP_HPM;
MatrixPrint(:,:,:,15) = I_MTF_GLP_HPM_H;
MatrixPrint(:,:,:,16) = I_MTF_GLP_HPM_R;
MatrixPrint(:,:,:,17) = I_MTF_GLP_CBD;
MatrixPrint(:,:,:,18) = I_C_MTF_GLP_CBD;
MatrixPrint(:,:,:,19) = I_MF;
MatrixPrint(:,:,:,20) = I_FE_HPM;
MatrixPrint(:,:,:,21) = I_SR_D;
MatrixPrint(:,:,:,22) = I_PWMBF;
MatrixPrint(:,:,:,23) = I_TV;
MatrixPrint(:,:,:,24) = I_RR;
MatrixPrint(:,:,:,25) = I_PNN;
MatrixPrint(:,:,:,26) = I_PNN_IDX;
MatrixPrint(:,:,:,27) = I_A_PNN;
MatrixPrint(:,:,:,28) = I_A_PNN_FT;
if size(I_MS,3) == 4
vect_index_RGB = [3,2,1];
else
vect_index_RGB = [5,3,2];
end
titleImages = algorithms;
addpath([pwd,'\Tools']);
figure, MP = showImagesAll(MatrixPrint,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0);
cd 'Outputs'
for ii = 1 : size(MP,4)
imwrite(MP(:,:,:,ii),sprintf('%s.png',algorithms{ii}));
end
imwrite(showPan(I_PAN,0,1,flag_cut_bounds,dim_cut),'PAN.png')
cd .. |
294coder/Efficient-MIF | 5,360 | Pansharpening_Hyper_SR_Matlab_Test_Package/Avg_RR_Assessment.tex | \begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0001&0.0025&0.0000&0.0016&0.0008&0.0264\\\hline
\textbf{BDSD-PC}&0.0001&0.0032&0.0000&0.0015&0.0008&0.0259\\\hline
\textbf{MTF-GLP-HPM-R}&0.0001&0.0017&0.0000&0.0010&0.0008&0.0278\\\hline
\textbf{MTF-GLP-FS}&0.0000&0.0016&0.0000&0.0012&0.0008&0.0277\\\hline
\textbf{TV}&0.0001&0.0018&0.0000&0.0008&0.0008&0.0279\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0001&0.0025&0.0000&0.0016&0.0008&0.0264\\\hline
\textbf{BDSD-PC}&0.0001&0.0032&0.0000&0.0015&0.0008&0.0259\\\hline
\textbf{MTF-GLP-HPM-R}&0.0001&0.0017&0.0000&0.0010&0.0008&0.0278\\\hline
\textbf{MTF-GLP-FS}&0.0000&0.0016&0.0000&0.0012&0.0008&0.0277\\\hline
\textbf{TV}&0.0001&0.0018&0.0000&0.0008&0.0008&0.0279\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline
\textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline
\textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline
\textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline
\textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline
\textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline
\textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline
\textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline
\textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline
\textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline
\textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline
\textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline
\textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0938&0.0341&0.1712&0.0388&0.7520&0.0572\\\hline
\textbf{BDSD-PC}&0.0960&0.0292&0.1652&0.0357&0.7553&0.0506\\\hline
\textbf{MTF-GLP-HPM-R}&0.0367&0.0133&0.1445&0.0341&0.8243&0.0393\\\hline
\textbf{MTF-GLP-FS}&0.0373&0.0140&0.1451&0.0339&0.8233&0.0398\\\hline
\textbf{TV}&0.1685&0.0480&0.1277&0.0460&0.7269&0.0724\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.1851&0.1848&0.1493&0.0937&0.7080&0.2157\\\hline
\textbf{BDSD-PC}&0.1505&0.1204&0.1464&0.1159&0.7375&0.1860\\\hline
\textbf{MTF-GLP-HPM-R}&0.0635&0.0575&0.1370&0.1179&0.8139&0.1507\\\hline
\textbf{MTF-GLP-FS}&0.0631&0.0579&0.1390&0.1191&0.8126&0.1524\\\hline
\textbf{TV}&0.1077&0.1145&0.1323&0.1361&0.7887&0.2078\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.2759&0.1309&0.1835&0.0861&0.5989&0.1528\\\hline
\textbf{BDSD-PC}&0.2171&0.0657&0.1929&0.1004&0.6373&0.1228\\\hline
\textbf{MTF-GLP-HPM-R}&0.0744&0.0372&0.1606&0.0787&0.7787&0.0918\\\hline
\textbf{MTF-GLP-FS}&0.0704&0.0325&0.1762&0.0777&0.7677&0.0908\\\hline
\textbf{TV}&0.0820&0.0366&0.1337&0.0835&0.7979&0.1043\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0878&0.0332&0.1712&0.0388&0.7569&0.0564\\\hline
\textbf{BDSD-PC}&0.0899&0.0286&0.1636&0.0362&0.7619&0.0505\\\hline
\textbf{MTF-GLP-HPM-R}&0.0365&0.0130&0.1574&0.0360&0.8121&0.0404\\\hline
\textbf{MTF-GLP-FS}&0.0371&0.0137&0.1583&0.0358&0.8107&0.0408\\\hline
\textbf{TV}&0.0696&0.0300&0.1448&0.0378&0.7957&0.0443\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0156&0.0040&0.0834&0.0101&0.9023&0.0106\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0540&0.0427&0.1045&0.0360&0.8479&0.0618\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0187&0.0045&0.0834&0.0101&0.8995&0.0108\\\hline
\end{tabular}
\begin{tabular}{|l|c|c|c|c|c|c|}
\hline
&\textbf{D_lambda}&\textbf{D_l-std}&\textbf{D_S}&\textbf{D_S-std}&\textbf{QNRI}&\textbf{QNRI-std}\\\hline
\textbf{BT-H}&0.0577&0.0213&0.1578&0.0295&0.7934&0.0262\\\hline
\end{tabular}
|
2977094657/DidaAPI | 1,997 | frontend/docs/api/users.md | # 获取用户信息
获取当前登录用户的详细信息。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/user/profile`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
无需参数,获取当前登录用户的信息。
## 响应格式
### 成功响应
```json
{
"etimestamp": null,
"username": "string",
"siteDomain": "dida365.com",
"createdCampaign": "string",
"createdDeviceInfo": null,
"filledPassword": true,
"accountDomain": "string",
"extenalId": "string",
"email": "string",
"verifiedEmail": true,
"fakedEmail": false,
"phone": null,
"name": "string",
"givenName": null,
"familyName": null,
"link": null,
"picture": "string",
"gender": "string",
"locale": "zh_CN",
"userCode": "string",
"verCode": null,
"verKey": null,
"externalId": "string",
"phoneWithoutCountryCode": null,
"displayName": "string"
}
```
## 响应字段说明
| 字段 | 类型 | 说明 |
|------|------|------|
| etimestamp | null | 时间戳 |
| username | string | 用户名(通常是邮箱) |
| siteDomain | string | 站点域名 |
| createdCampaign | string | 创建活动 |
| createdDeviceInfo | null | 创建设备信息 |
| filledPassword | boolean | 是否已设置密码 |
| accountDomain | string | 账户域名 |
| extenalId | string | 外部ID |
| email | string | 邮箱地址 |
| verifiedEmail | boolean | 邮箱是否已验证 |
| fakedEmail | boolean | 是否为虚假邮箱 |
| phone | string/null | 手机号 |
| name | string | 姓名 |
| givenName | string/null | 名 |
| familyName | string/null | 姓 |
| link | string/null | 链接 |
| picture | string | 头像URL |
| gender | string | 性别(0-未知,1-男,2-女) |
| locale | string | 语言设置 |
| userCode | string | 用户代码 |
| verCode | string/null | 验证码 |
| verKey | string/null | 验证密钥 |
| externalId | string | 外部标识 |
| phoneWithoutCountryCode | string/null | 不含国家代码的手机号 |
| displayName | string | 显示名称 |
## 使用说明
1. 确保已完成认证获取会话
2. 直接调用接口即可获取当前用户信息
3. 返回的数据包含用户的完整个人资料信息
4. 头像URL可直接用于显示用户头像
5. 用户代码(userCode)是用户的唯一标识
## 应用场景
- **用户资料展示**: 在应用中显示用户的基本信息
- **头像显示**: 获取用户头像URL用于界面展示
- **用户身份验证**: 确认当前登录用户的身份
- **个人设置**: 获取用户的语言、性别等设置信息
- **账户管理**: 显示账户相关信息如邮箱验证状态
|
2977094657/DidaAPI | 2,232 | frontend/docs/api/projects.md | # 获取清单列表
获取用户的所有清单列表。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/projects`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
无需参数,使用当前认证会话。
## 响应格式
### 成功响应
```json
[
{
"id": "string",
"name": "string",
"isOwner": true,
"color": "string",
"sortOrder": 0,
"sortOption": {
"groupBy": "string",
"orderBy": "string"
},
"sortType": "string",
"userCount": 0,
"etag": "string",
"modifiedTime": "string",
"inAll": true,
"showType": null,
"muted": true,
"reminderType": null,
"closed": null,
"transferred": null,
"groupId": null,
"viewMode": "string",
"notificationOptions": [
"string"
],
"teamId": null,
"permission": "string",
"kind": "string",
"timeline": {
"range": null,
"sortType": "string",
"sortOption": {
"groupBy": "string",
"orderBy": "string"
}
},
"needAudit": true,
"barcodeNeedAudit": true,
"openToTeam": true,
"teamMemberPermission": null,
"source": 0
}
]
```
## 响应字段说明
| 字段 | 类型 | 说明 |
|------|------|------|
| `id` | string | 清单唯一标识符 |
| `name` | string | 清单名称 |
| `isOwner` | boolean | 是否为拥有者 |
| `color` | string | 清单颜色 |
| `sortOrder` | number | 自定义排序值 |
| `sortOption` | object | 排序选项配置 |
| `sortType` | string | 排序类型 |
| `userCount` | number | 用户数量 |
| `etag` | string | 实体标签 |
| `modifiedTime` | string | 修改时间 |
| `inAll` | boolean | 是否在全部清单中显示 |
| `showType` | any | 显示类型 |
| `muted` | boolean | 是否静音 |
| `reminderType` | any | 提醒类型 |
| `closed` | any | 关闭状态 |
| `transferred` | any | 转移状态 |
| `groupId` | string/null | 分组ID |
| `viewMode` | string | 视图模式 |
| `notificationOptions` | array | 通知选项 |
| `teamId` | string/null | 团队ID |
| `permission` | string | 权限类型 |
| `kind` | string | 清单类型 |
| `timeline` | object | 时间线配置 |
| `needAudit` | boolean | 是否需要审核 |
| `barcodeNeedAudit` | boolean | 条码是否需要审核 |
| `openToTeam` | boolean | 是否对团队开放 |
| `teamMemberPermission` | any | 团队成员权限 |
| `source` | number | 来源标识 |
|
2977094657/DidaAPI | 1,715 | frontend/docs/api/index.md | # 滴答清单API文档
本文档详细描述了滴答清单的原始API接口,帮助开发者了解如何直接调用滴答清单的服务。
## 接口分类
### 认证相关
- [微信登录流程](./auth/wechat-login-flow.md) - 完整的微信扫码登录流程
- [获取微信二维码](./auth/get-wechat-qrcode.md) - 获取微信登录二维码
- [轮询登录状态](./auth/poll-login-status.md) - 检查二维码扫码状态
- [验证微信登录](./auth/validate-wechat-login.md) - 验证微信登录并获取令牌
- [微信登录回调处理](./auth/wechat-callback.md) - 处理微信扫码后的回调,提取code参数
- [密码登录](./auth/password-login.md) - 使用用户名和密码进行登录
### 任务管理
- [获取所有任务](./tasks/get-all-tasks.md) - 获取用户的所有任务列表
- [获取已完成任务](./tasks/get-completed-tasks.md) - 获取已完成/已放弃的任务列表,支持分页
- [获取垃圾桶任务](./tasks/get-trash-tasks.md) - 获取垃圾桶中的任务列表
- [获取任务统计](./tasks/get-tasks-summary.md) - 获取任务的统计信息
### 清单管理
- [获取清单列表](./projects.md) - 获取用户的所有清单列表
### 统计分析
- [获取用户排名统计](./statistics.md) - 获取用户在滴答清单中的排名和基本统计信息
- [获取通用统计信息](./statistics/general-statistics.md) - 获取概览、成就值、趋势等通用统计信息
- [获取任务统计信息](./statistics/task-statistics.md) - 获取指定日期范围内的任务统计信息
### 番茄专注
- [获取番茄专注概览](./pomodoros.md) - 获取番茄专注的概览统计信息
### 正计时专注
- [获取专注记录时间线](./pomodoros/focus-timeline.md) - 获取专注记录的时间线数据,支持分页
- [获取专注详情分布](./pomodoros/focus-distribution.md) - 获取指定日期范围内的专注时长分布统计
- [获取专注趋势热力图](./pomodoros/focus-heatmap.md) - 获取指定日期范围内的专注趋势热力图数据
- [获取专注时间按小时分布](./pomodoros/focus-hour-distribution.md) - 获取指定日期范围内按小时分布的专注时间统计
- [获取专注时间分布](./pomodoros/focus-time-distribution.md) - 获取指定日期范围内按时间段分布的专注数据
### 习惯管理
- [获取所有习惯](./habits.md) - 获取当前用户的所有习惯列表
- [获取本周习惯打卡统计](./habits/week-current-statistics.md) - 获取本周的习惯打卡统计信息
- [导出习惯数据](./habits/export-habits.md) - 导出用户的习惯数据为Excel文件
### 用户信息
- [获取用户信息](./users.md) - 获取当前登录用户的详细信息
### 自定义接口
- [导出任务到Excel](./custom/export-tasks-excel.md) - 导出所有任务到Excel文件,包含全部任务、已完成任务、垃圾桶任务三个工作表
- [导出专注记录到Excel](./custom/export-focus-excel.md) - 导出所有专注记录到Excel文件,包含完整的专注时间线数据
|
294coder/Efficient-MIF | 30,245 | Pansharpening_Hyper_SR_Matlab_Test_Package/Demo1_Reduced_Resolution_WV3_MultiExm.m | %This is a demo to run fusion algorithms on Reduced Resolution
% LJ Deng(UESTC)
% 2022-06-02
clear; close all;
%% =======load directors========
% Tools
addpath([pwd,'/Tools']);
% Select algorithms to run
algorithms = {'GT'};%'PNN'
location1 = [40 60 4 43]; % Location of zoom in
location2 = [];
%sensor = 'WV3';
%% =======read Multiple TestData_wv3.h5 (four 512x512 WV3 simulated data)========
file_test = '/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5';
disp(file_test)
gt_multiExm_tmp = h5read(file_test,'/gt'); % WxHxCxN=1x2x3x4
gt_multiExm = permute(gt_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
ms_multiExm_tmp = h5read(file_test,'/ms'); % WxHxCxN=1x2x3x4
ms_multiExm = permute(ms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
lms_multiExm_tmp = h5read(file_test,'/lms'); % WxHxCxN=1x2x3x4
lms_multiExm = permute(lms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
pan_multiExm_tmp = h5read(file_test,'/pan'); % WxHxCxN=1x2x3x4
pan_multiExm = permute(pan_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
data_name = '3_EPS/WV3/multi/'; % director to save EPS figures
%% ==========Read each Data====================
exm_num = size(ms_multiExm, 1);
for i = 1 : exm_num % i = 1 or 2 ...
%% read each data
HRMS_tmp = gt_multiExm(i, :, :, :); % I_GT
I_GT = squeeze(HRMS_tmp);
LRMS_tmp = ms_multiExm(i, :, :, :); % I_MS_LR
I_MS_LR = squeeze(LRMS_tmp);
LMS_tmp = lms_multiExm(i, :, :, :); % I_MS
I_MS = squeeze(LMS_tmp);
PAN_tmp = pan_multiExm(i, :, :, :); % I_PAN
I_PAN = squeeze(PAN_tmp);
%% Initialization of the Matrix of Results
NumIndexes = 5;
MatrixResults = zeros(numel(algorithms),NumIndexes);
alg = 0;
%% load Indexes for WV3_RR
sensor = 'GF2';
% disp(sensor)
Qblocks_size = 32;
bicubic = 0;% Interpolator
flag_cut_bounds = 1;% Cut Final Image
dim_cut = 30;% Cut Final Image
thvalues = 0;% Threshold values out of dynamic range
printEPS = 0;% Print Eps
ratio = 4;% Resize Factor
L = 11;% Radiometric Resolution
%% show I_MS_LR, I_GT, PAN Imgs:
% showImage8_zoomin(I_MS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% showPan(I_PAN,printEPS,2,flag_cut_bounds,dim_cut);
% pause(2);print('-deps', strcat(data_name, num2str(i-1), '_pan', '.eps'))
%% ======GT ===================
if ismember('GT',algorithms)
alg = alg + 1;
[Q_avg_GT, SAM_GT, ERGAS_GT, SCC_GT, Q_GT] = indexes_evaluation(I_GT,I_MS,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_GT,Q_avg_GT,SAM_GT,ERGAS_GT,SCC_GT];
MatrixImage(:,:,:,alg) = I_GT;
Q_avg_GT_multiexm(i) = Q_avg_GT;
SAM_GT_multiexm(i) = SAM_GT;
ERGAS_GT_multiexm(i) = ERGAS_GT;
SCC_GT_multiexm(i) = SCC_GT;
Q_GT_multiexm(i) = Q_GT;
% showImage8_zoomin(I_GT,printEPS,2,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_gt', '.eps'))
end
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%% CS-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%%
%% ====== 1) BT-H Method ======
if ismember('BT-H',algorithms)
alg = alg + 1;
cd BT-H
t2=tic;
I_BT_H = BroveyRegHazeMin(I_MS,I_PAN,ratio);
time_BT_H = toc(t2);
fprintf('Elaboration time BT-H: %.2f [sec]\n',time_BT_H);
cd ..
%%% Quality indexes computation
[Q_avg_BT_H, SAM_BT_H, ERGAS_BT_H, SCC_BT_H, Q_BT_H] = indexes_evaluation(I_BT_H,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_BT_H,Q_avg_BT_H,SAM_BT_H,ERGAS_BT_H,SCC_BT_H];
MatrixImage(:,:,:,alg) = I_BT_H;
Q_avg_BT_H_multiexm(i) = Q_avg_BT_H;
SAM_BT_H_multiexm(i) = SAM_BT_H;
ERGAS_BT_H_multiexm(i) = ERGAS_BT_H;
SCC_BT_H_multiexm(i) = SCC_BT_H;
Q_BT_H_multiexm(i) = Q_BT_H;
% showImage8_zoomin(I_BT_H,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bth.eps'))
end
%% ====== 2) BDSD-PC Method ======
if ismember('BDSD-PC',algorithms)
alg = alg + 1;
cd BDSD
t2=tic;
I_BDSD_PC = BDSD_PC(I_MS,I_PAN,ratio,sensor);
time_BDSD_PC = toc(t2);
fprintf('Elaboration time BDSD-PC: %.2f [sec]\n',time_BDSD_PC);
cd ..
[Q_avg_BDSD_PC, SAM_BDSD_PC, ERGAS_BDSD_PC, SCC_BDSD_PC, Q_BDSD_PC] = indexes_evaluation(I_BDSD_PC,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_BDSD_PC,Q_avg_BDSD_PC,SAM_BDSD_PC,ERGAS_BDSD_PC,SCC_BDSD_PC];
MatrixImage(:,:,:,alg) = I_BDSD_PC;
Q_avg_BDSD_PC_multiexm(i) = Q_avg_BDSD_PC;
SAM_BDSD_PC_multiexm(i) = SAM_BDSD_PC;
ERGAS_BDSD_PC_multiexm(i) = ERGAS_BDSD_PC;
SCC_BDSD_PC_multiexm(i) = SCC_BDSD_PC;
Q_BDSD_PC_multiexm(i) = Q_BDSD_PC;
% showImage8_zoomin(I_BDSD_PC,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bdsd_pc.eps'))
end
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%% MRA-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%%
%% ====== 1) MTF-GLP-HPM-R Method ======
if ismember('MTF-GLP-HPM-R',algorithms)
alg = alg + 1;
cd GLP
t2=tic;
I_MTF_GLP_HPM_R = MTF_GLP_HPM_R(I_MS,I_PAN,sensor,ratio);
time_MTF_GLP_HPM_R = toc(t2);
fprintf('Elaboration time MTF-GLP: %.2f [sec]\n',time_MTF_GLP_HPM_R);
cd ..
[Q_avg_MTF_GLP_HPM_R, SAM_MTF_GLP_HPM_R, ERGAS_MTF_GLP_HPM_R, SCC_MTF_GLP_HPM_R, Q_MTF_GLP_HPM_R] = indexes_evaluation(I_MTF_GLP_HPM_R,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_MTF_GLP_HPM_R,Q_avg_MTF_GLP_HPM_R,SAM_MTF_GLP_HPM_R,ERGAS_MTF_GLP_HPM_R,SCC_MTF_GLP_HPM_R];
MatrixImage(:,:,:,alg) = I_MTF_GLP_HPM_R;
Q_avg_MTF_GLP_HPM_R_multiexm(i) = Q_avg_MTF_GLP_HPM_R;
SAM_MTF_GLP_HPM_R_multiexm(i) = SAM_MTF_GLP_HPM_R;
ERGAS_MTF_GLP_HPM_R_multiexm(i) = ERGAS_MTF_GLP_HPM_R;
SCC_MTF_GLP_HPM_R_multiexm(i) = SCC_MTF_GLP_HPM_R;
Q_MTF_GLP_HPM_R_multiexm(i) = Q_MTF_GLP_HPM_R;
% showImage8_zoomin(I_MTF_GLP_HPM_R,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglp_hpm_r.eps'))
end
%% ====== 2) MTF-GLP-FS Method ======
if ismember('MTF-GLP-FS',algorithms)
alg = alg + 1;
cd GLP
t2=tic;
I_MTF_GLP_FS = MTF_GLP_FS(I_MS,I_PAN,sensor,ratio);
time_MTF_GLP_FS = toc(t2);
fprintf('Elaboration time MTF-GLP-FS: %.2f [sec]\n',time_MTF_GLP_FS);
cd ..
[Q_avg_MTF_GLP_FS, SAM_MTF_GLP_FS, ERGAS_MTF_GLP_FS, SCC_MTF_GLP_FS, Q_MTF_GLP_FS] = indexes_evaluation(I_MTF_GLP_FS,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_MTF_GLP_FS,Q_avg_MTF_GLP_FS,SAM_MTF_GLP_FS,ERGAS_MTF_GLP_FS,SCC_MTF_GLP_FS];
MatrixImage(:,:,:,alg) = I_MTF_GLP_FS;
Q_avg_MTF_GLP_FS_multiexm(i) = Q_avg_MTF_GLP_FS;
SAM_MTF_GLP_FS_multiexm(i) = SAM_MTF_GLP_FS;
ERGAS_MTF_GLP_FS_multiexm(i) = ERGAS_MTF_GLP_FS;
SCC_MTF_GLP_FS_multiexm(i) = SCC_MTF_GLP_FS;
Q_MTF_GLP_FS_multiexm(i) = Q_MTF_GLP_FS;
% showImage8_zoomin(I_MTF_GLP_FS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglpfs.eps'))
end
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%% VO-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%%
%% ====== 1) TV Method ======
if ismember('TV',algorithms)
alg = alg + 1;
%%%%%%%%%%%%%%%%%%%%%%%%%% Parameters setting %%%%%%%%%%%%%%%%%%%%%%%%%%%%%
switch sensor
case 'IKONOS'
w=[0.1091 0.2127 0.2928 0.3854];
c = 8;
alpha=1.064;
maxiter=10;
lambda = 0.47106;
case {'GeoEye1','WV4'}
w=[0.1552, 0.3959, 0.2902, 0.1587];
c = 8;
alpha=0.75;
maxiter=50;
lambda = 157.8954;
case 'WV3'
w=[0.0657 0.1012 0.1537 0.1473 0.1245 0.1545 0.1338 0.1192];
c = 8;
alpha=0.75;
maxiter=50;
lambda = 1.0000e-03;
end
cd TV
t2 = tic;
I_TV = TV_pansharpen(I_MS_LR,I_PAN,alpha,lambda,c,maxiter,w);
time_TV = toc(t2);
fprintf('Elaboration time TV: %.2f [sec]\n',time_TV);
cd ..
[Q_avg_TV, SAM_TV, ERGAS_TV, SCC_TV, Q_TV] = indexes_evaluation(I_TV,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
MatrixResults(alg,:) = [Q_TV,Q_avg_TV,SAM_TV,ERGAS_TV,SCC_TV];
MatrixImage(:,:,:,alg) = I_TV;
Q_avg_TV_multiexm(i) = Q_avg_TV;
SAM_TV_multiexm(i) = SAM_TV;
ERGAS_TV_multiexm(i) = ERGAS_TV;
SCC_TV_multiexm(i) = SCC_TV;
Q_TV_multiexm(i) = Q_TV;
% showImage8_zoomin(I_TV,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_tv.eps'))
end
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%% DL-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%%
%% ====== 1) PNN Method ======
% load(strcat('2_DL_Result/WV3_Reduced/PNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PNN
% I_pnn = double(sr);
%
% if ismember('PNN',algorithms)
% alg = alg + 1;
% [Q_avg_pnn, SAM_pnn, ERGAS_pnn, SCC_pnn, Q_pnn] = indexes_evaluation(I_pnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_pnn,Q_avg_pnn,SAM_pnn,ERGAS_pnn,SCC_pnn];
% MatrixImage(:,:,:,alg) = I_pnn;
%
% Q_avg_pnn_multiexm(i) = Q_avg_pnn;
% SAM_pnn_multiexm(i) = SAM_pnn;
% ERGAS_pnn_multiexm(i) = ERGAS_pnn;
% SCC_pnn_multiexm(i) = SCC_pnn;
% Q_pnn_multiexm(i) = Q_pnn;
%
% showImage8_zoomin(I_pnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pnn.eps'))
% end
% %% ====== 2) PanNet Method ======
%
% load(strcat('2_DL_Result/WV3_Reduced/PanNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PanNet
% I_pannet = double(sr);
%
% if ismember('PanNet',algorithms)
% alg = alg + 1;
% [Q_avg_pannet, SAM_pannet, ERGAS_pannet, SCC_pannet, Q_pannet] = indexes_evaluation(I_pannet,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_pannet,Q_avg_pannet,SAM_pannet,ERGAS_pannet,SCC_pannet];
% MatrixImage(:,:,:,alg) = I_pannet;
%
%
% Q_avg_pannet_multiexm(i) = Q_avg_pannet;
% SAM_pannet_multiexm(i) = SAM_pannet;
% ERGAS_pannet_multiexm(i) = ERGAS_pannet;
% SCC_pannet_multiexm(i) = SCC_pannet;
% Q_pannet_multiexm(i) = Q_pannet;
%
% showImage8_zoomin(I_pannet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pannet.eps'))
% end
%% ====== 3) DiCNN Method ======
% load(strcat('2_DL_Result/WV3_Reduced/DiCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for DiCNN
% I_dicnn = double(sr);
%
% if ismember('DiCNN',algorithms)
% alg = alg + 1;
% [Q_avg_dicnn, SAM_dicnn, ERGAS_dicnn, SCC_dicnn, Q_dicnn] = indexes_evaluation(I_dicnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_dicnn,Q_avg_dicnn,SAM_dicnn,ERGAS_dicnn,SCC_dicnn];
% MatrixImage(:,:,:,alg) = I_dicnn;
%
% Q_avg_dicnn_multiexm(i) = Q_avg_dicnn;
% SAM_dicnn_multiexm(i) = SAM_dicnn;
% ERGAS_dicnn_multiexm(i) = ERGAS_dicnn;
% SCC_dicnn_multiexm(i) = SCC_dicnn;
% Q_dicnn_multiexm(i) = Q_dicnn;
%
% showImage8_zoomin(I_dicnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_dicnn.eps'))
% end
%
% %% ====== 4) MSDCNN Method ======
% load(strcat('2_DL_Result/WV3_Reduced/MSDCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for MSDCNN
% I_msdcnn = double(sr);
%
% if ismember('MSDCNN',algorithms)
% alg = alg + 1;
% [Q_avg_msdcnn, SAM_msdcnn, ERGAS_msdcnn, SCC_msdcnn, Q_msdcnn] = indexes_evaluation(I_msdcnn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_msdcnn,Q_avg_msdcnn,SAM_msdcnn,ERGAS_msdcnn,SCC_msdcnn];
% MatrixImage(:,:,:,alg) = I_msdcnn;
%
% Q_avg_msdcnn_multiexm(i) = Q_avg_msdcnn;
% SAM_msdcnn_multiexm(i) = SAM_msdcnn;
% ERGAS_msdcnn_multiexm(i) = ERGAS_msdcnn;
% SCC_msdcnn_multiexm(i) = SCC_msdcnn;
% Q_msdcnn_multiexm(i) = Q_msdcnn;
%
% showImage8_zoomin(I_msdcnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_msdcnn.eps'))
% end
% %% ====== 5) BDPN Method ======
% load(strcat('2_DL_Result/WV3_Reduced/BDPN/results/output_mulExm_' , num2str(i-1), '.mat')) % load i-th image for BDPN
% I_bdpn = double(sr);
%
% if ismember('BDPN',algorithms)
% alg = alg + 1;
% [Q_avg_bdpn, SAM_bdpn, ERGAS_bdpn, SCC_bdpn, Q_bdpn] = indexes_evaluation(I_bdpn,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_bdpn,Q_avg_bdpn,SAM_bdpn,ERGAS_bdpn,SCC_bdpn];
% MatrixImage(:,:,:,alg) = I_bdpn;
%
% Q_avg_bdpn_multiexm(i) = Q_avg_bdpn;
% SAM_bdpn_multiexm(i) = SAM_bdpn;
% ERGAS_bdpn_multiexm(i) = ERGAS_bdpn;
% SCC_bdpn_multiexm(i) = SCC_bdpn;
% Q_bdpn_multiexm(i) = Q_bdpn;
%
% showImage8_zoomin(I_bdpn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_bdpn.eps'))
% end
%
% %% ====== 6) FusionNet Method ======
% load(strcat('2_DL_Result/WV3_Reduced/FusionNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for FusionNet
% I_fusionnet = double(sr);
%
% if ismember('FusionNet',algorithms)
% alg = alg + 1;
% [Q_avg_fusionnet, SAM_fusionnet, ERGAS_fusionnet, SCC_fusionnet, Q_fusionnet] = indexes_evaluation(I_fusionnet,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_fusionnet,Q_avg_fusionnet,SAM_fusionnet,ERGAS_fusionnet,SCC_fusionnet];
% MatrixImage(:,:,:,alg) = I_fusionnet;
%
% Q_avg_fusionnet_multiexm(i) = Q_avg_fusionnet;
% SAM_fusionnet_multiexm(i) = SAM_fusionnet;
% ERGAS_fusionnet_multiexm(i) = ERGAS_fusionnet;
% SCC_fusionnet_multiexm(i) = SCC_fusionnet;
% Q_fusionnet_multiexm(i) = Q_fusionnet;
%
% showImage8_zoomin(I_fusionnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_fusionnet.eps'))
% end
%% ====== 7) LAGConv Method ======
% load(strcat('2_DL_Result/WV3_Reduced/LAGConv/results/output_mulExm_',num2str(i-1), '.mat')) % load i-th image for LAGConv
% I_lagnet = double(sr);
%
% if ismember('LAGConv',algorithms)
% alg = alg + 1;
% [Q_avg_lagnet, SAM_lagnet, ERGAS_lagnet, SCC_lagnet, Q_lagnet] = indexes_evaluation(I_lagnet,I_GT,ratio,L,Qblocks_size,flag_cut_bounds,dim_cut,thvalues);
% MatrixResults(alg,:) = [Q_lagnet,Q_avg_lagnet,SAM_lagnet,ERGAS_lagnet,SCC_lagnet];
% MatrixImage(:,:,:,alg) = I_lagnet;
%
% Q_avg_lagnet_multiexm(i) = Q_avg_lagnet;
% SAM_lagnet_multiexm(i) = SAM_lagnet;
% ERGAS_lagnet_multiexm(i) = ERGAS_lagnet;
% SCC_lagnet_multiexm(i) = SCC_lagnet;
% Q_lagnet_multiexm(i) = Q_lagnet;
%
% showImage8_zoomin(I_lagnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L,location1,location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_lagnet.eps'))
% end
end
%% Print in LATEX
%matrix2latex(MatrixResults(:,[1,3,4]),'RR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'Q2n'},{'SAM'},{'ERGAS'}],'alignment','c','format', '%.4f');
%% View All
if size(I_GT,3) == 4
vect_index_RGB = [3,2,1];
else
vect_index_RGB = [5,3,2];
end
titleImages = algorithms;
figure, showImagesAll(MatrixImage,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0);
%% ======Display the final average performance =======
% GT: average Q_avg
avg_Q_GT_multiexm = mean(Q_GT_multiexm);
std_Q_GT_multiexm = std(Q_GT_multiexm);
avg_Q_avg_GT_multiexm = mean(Q_avg_GT_multiexm);
std_Q_avg_GT_multiexm = std(Q_avg_GT_multiexm);
avg_SAM_GT_multiexm = mean(SAM_GT_multiexm);
std_SAM_GT_multiexm = std(SAM_GT_multiexm);
avg_ERGAS_GT_multiexm = mean(ERGAS_GT_multiexm);
std_ERGAS_GT_multiexm = std(ERGAS_GT_multiexm);
avg_SCC_GT_multiexm = mean(SCC_GT_multiexm);
std_SCC_GT_multiexm = std(SCC_GT_multiexm);
Avg_MatrixResults(1,:) = [avg_Q_GT_multiexm, std_Q_GT_multiexm, avg_Q_avg_GT_multiexm, std_Q_avg_GT_multiexm, ...
avg_SAM_GT_multiexm, std_SAM_GT_multiexm, avg_ERGAS_GT_multiexm, std_ERGAS_GT_multiexm,...
avg_SCC_GT_multiexm, std_SCC_GT_multiexm];
% BT_H: average Q_avg
% avg_Q_BT_H_multiexm = mean(Q_BT_H_multiexm);
% std_Q_BT_H_multiexm = std(Q_BT_H_multiexm);
% avg_Q_avg_BT_H_multiexm = mean(Q_avg_BT_H_multiexm);
% std_Q_avg_BT_H_multiexm = std(Q_avg_BT_H_multiexm);
% avg_SAM_BT_H_multiexm = mean(SAM_BT_H_multiexm);
% std_SAM_BT_H_multiexm = std(SAM_BT_H_multiexm);
% avg_ERGAS_BT_H_multiexm = mean(ERGAS_BT_H_multiexm);
% std_ERGAS_BT_H_multiexm = std(ERGAS_BT_H_multiexm);
% avg_SCC_BT_H_multiexm = mean(SCC_BT_H_multiexm);
% std_SCC_BT_H_multiexm = std(SCC_BT_H_multiexm);
% Avg_MatrixResults(2,:) = [avg_Q_BT_H_multiexm, std_Q_BT_H_multiexm, avg_Q_avg_BT_H_multiexm, std_Q_avg_BT_H_multiexm, ...
% avg_SAM_BT_H_multiexm, std_SAM_BT_H_multiexm, avg_ERGAS_BT_H_multiexm, std_ERGAS_BT_H_multiexm,...
% avg_SCC_BT_H_multiexm, std_SCC_BT_H_multiexm];
% % BDSD: average Q_avg
% avg_Q_BDSD_PC_multiexm = mean(Q_BDSD_PC_multiexm);
% std_Q_BDSD_PC_multiexm = std(Q_BDSD_PC_multiexm);
% avg_Q_avg_BDSD_PC_multiexm = mean(Q_avg_BDSD_PC_multiexm);
% std_Q_avg_BDSD_PC_multiexm = std(Q_avg_BDSD_PC_multiexm);
% avg_SAM_BDSD_PC_multiexm = mean(SAM_BDSD_PC_multiexm);
% std_SAM_BDSD_PC_multiexm = std(SAM_BDSD_PC_multiexm);
% avg_ERGAS_BDSD_PC_multiexm = mean(ERGAS_BDSD_PC_multiexm);
% std_ERGAS_BDSD_PC_multiexm = std(ERGAS_BDSD_PC_multiexm);
% avg_SCC_BDSD_PC_multiexm = mean(SCC_BDSD_PC_multiexm);
% std_SCC_BDSD_PC_multiexm = std(SCC_BDSD_PC_multiexm);
% Avg_MatrixResults(3,:) = [avg_Q_BDSD_PC_multiexm, std_Q_BDSD_PC_multiexm, avg_Q_avg_BDSD_PC_multiexm, std_Q_avg_BDSD_PC_multiexm, ...
% avg_SAM_BDSD_PC_multiexm, std_SAM_BDSD_PC_multiexm, avg_ERGAS_BDSD_PC_multiexm, std_ERGAS_BDSD_PC_multiexm,...
% avg_SCC_BDSD_PC_multiexm, std_SCC_BDSD_PC_multiexm];
% % MTF_GLP: average Q_avg
% avg_Q_MTF_GLP_HPM_R_multiexm = mean(Q_MTF_GLP_HPM_R_multiexm);
% std_Q_MTF_GLP_HPM_R_multiexm = std(Q_MTF_GLP_HPM_R_multiexm);
% avg_Q_avg_MTF_GLP_HPM_R_multiexm = mean(Q_avg_MTF_GLP_HPM_R_multiexm);
% std_Q_avg_MTF_GLP_HPM_R_multiexm = std(Q_avg_MTF_GLP_HPM_R_multiexm);
% avg_SAM_MTF_GLP_HPM_R_multiexm = mean(SAM_MTF_GLP_HPM_R_multiexm);
% std_SAM_MTF_GLP_HPM_R_multiexm = std(SAM_MTF_GLP_HPM_R_multiexm);
% avg_ERGAS_MTF_GLP_HPM_R_multiexm = mean(ERGAS_MTF_GLP_HPM_R_multiexm);
% std_ERGAS_MTF_GLP_HPM_R_multiexm = std(ERGAS_MTF_GLP_HPM_R_multiexm);
% avg_SCC_MTF_GLP_HPM_R_multiexm = mean(SCC_MTF_GLP_HPM_R_multiexm);
% std_SCC_MTF_GLP_HPM_R_multiexm = std(SCC_MTF_GLP_HPM_R_multiexm);
% Avg_MatrixResults(4,:) = [avg_Q_MTF_GLP_HPM_R_multiexm, std_Q_MTF_GLP_HPM_R_multiexm, avg_Q_avg_MTF_GLP_HPM_R_multiexm, std_Q_avg_MTF_GLP_HPM_R_multiexm, ...
% avg_SAM_MTF_GLP_HPM_R_multiexm, std_SAM_MTF_GLP_HPM_R_multiexm, avg_ERGAS_MTF_GLP_HPM_R_multiexm, std_ERGAS_MTF_GLP_HPM_R_multiexm,...
% avg_SCC_MTF_GLP_HPM_R_multiexm, std_SCC_MTF_GLP_HPM_R_multiexm];
% % MTF_GLP_FS: average Q_avg
% avg_Q_MTF_GLP_FS_multiexm = mean(Q_MTF_GLP_FS_multiexm);
% std_Q_MTF_GLP_FS_multiexm = std(Q_MTF_GLP_FS_multiexm);
% avg_Q_avg_MTF_GLP_FS_multiexm = mean(Q_avg_MTF_GLP_FS_multiexm);
% std_Q_avg_MTF_GLP_FS_multiexm = std(Q_avg_MTF_GLP_FS_multiexm);
% avg_SAM_MTF_GLP_FS_multiexm = mean(SAM_MTF_GLP_FS_multiexm);
% std_SAM_MTF_GLP_FS_multiexm = std(SAM_MTF_GLP_FS_multiexm);
% avg_ERGAS_MTF_GLP_FS_multiexm = mean(ERGAS_MTF_GLP_FS_multiexm);
% std_ERGAS_MTF_GLP_FS_multiexm = std(ERGAS_MTF_GLP_FS_multiexm);
% avg_SCC_MTF_GLP_FS_multiexm = mean(SCC_MTF_GLP_FS_multiexm);
% std_SCC_MTF_GLP_FS_multiexm = std(SCC_MTF_GLP_FS_multiexm);
% Avg_MatrixResults(5,:) = [avg_Q_MTF_GLP_FS_multiexm, std_Q_MTF_GLP_FS_multiexm, avg_Q_avg_MTF_GLP_FS_multiexm, std_Q_avg_MTF_GLP_FS_multiexm, ...
% avg_SAM_MTF_GLP_FS_multiexm, std_SAM_MTF_GLP_FS_multiexm, avg_ERGAS_MTF_GLP_FS_multiexm, std_ERGAS_MTF_GLP_FS_multiexm,...
% avg_SCC_MTF_GLP_FS_multiexm, std_SCC_MTF_GLP_FS_multiexm];
% % TV: average Q_avg
% avg_Q_TV_multiexm = mean(Q_TV_multiexm);
% std_Q_TV_multiexm = std(Q_TV_multiexm);
% avg_Q_avg_TV_multiexm = mean(Q_avg_TV_multiexm);
% std_Q_avg_TV_multiexm = std(Q_avg_TV_multiexm);
% avg_SAM_TV_multiexm = mean(SAM_TV_multiexm);
% std_SAM_TV_multiexm = std(SAM_TV_multiexm);
% avg_ERGAS_TV_multiexm = mean(ERGAS_TV_multiexm);
% std_ERGAS_TV_multiexm = std(ERGAS_TV_multiexm);
% avg_SCC_TV_multiexm = mean(SCC_TV_multiexm);
% std_SCC_TV_multiexm = std(SCC_TV_multiexm);
% Avg_MatrixResults(6,:) = [avg_Q_TV_multiexm, std_Q_TV_multiexm, avg_Q_avg_TV_multiexm, std_Q_avg_TV_multiexm, ...
% avg_SAM_TV_multiexm, std_SAM_TV_multiexm, avg_ERGAS_TV_multiexm, std_ERGAS_TV_multiexm,...
% avg_SCC_TV_multiexm, std_SCC_TV_multiexm];
% pnn: average Q_avg
%avg_Q_pnn_multiexm = mean(Q_pnn_multiexm);
%std_Q_pnn_multiexm = std(Q_pnn_multiexm);
%
%avg_Q_avg_pnn_multiexm = mean(Q_avg_pnn_multiexm);
%std_Q_avg_pnn_multiexm = std(Q_avg_pnn_multiexm);
%
%avg_SAM_pnn_multiexm = mean(SAM_pnn_multiexm);
%std_SAM_pnn_multiexm = std(SAM_pnn_multiexm);
%
%avg_ERGAS_pnn_multiexm = mean(ERGAS_pnn_multiexm);
%std_ERGAS_pnn_multiexm = std(ERGAS_pnn_multiexm);
%
%avg_SCC_pnn_multiexm = mean(SCC_pnn_multiexm);
%std_SCC_pnn_multiexm = std(SCC_pnn_multiexm);
%
%Avg_MatrixResults(7,:) = [avg_Q_pnn_multiexm, std_Q_pnn_multiexm, avg_Q_avg_pnn_multiexm, std_Q_avg_pnn_multiexm, ...
% avg_SAM_pnn_multiexm, std_SAM_pnn_multiexm, avg_ERGAS_pnn_multiexm, std_ERGAS_pnn_multiexm,...
% avg_SCC_pnn_multiexm, std_SCC_pnn_multiexm];
% pannet: average Q_avg
%avg_Q_pannet_multiexm = mean(Q_pannet_multiexm);
%std_Q_pannet_multiexm = std(Q_pannet_multiexm);
%
%avg_Q_avg_pannet_multiexm = mean(Q_avg_pannet_multiexm);
%std_Q_avg_pannet_multiexm = std(Q_avg_pannet_multiexm);
%
%avg_SAM_pannet_multiexm = mean(SAM_pannet_multiexm);
%std_SAM_pannet_multiexm = std(SAM_pannet_multiexm);
%
%avg_ERGAS_pannet_multiexm = mean(ERGAS_pannet_multiexm);
%std_ERGAS_pannet_multiexm = std(ERGAS_pannet_multiexm);
%
%avg_SCC_pannet_multiexm = mean(SCC_pannet_multiexm);
%std_SCC_pannet_multiexm = std(SCC_pannet_multiexm);
%
%Avg_MatrixResults(8,:) = [avg_Q_pannet_multiexm, std_Q_pannet_multiexm, avg_Q_avg_pannet_multiexm, std_Q_avg_pannet_multiexm, ...
% avg_SAM_pannet_multiexm, std_SAM_pannet_multiexm, avg_ERGAS_pannet_multiexm, std_ERGAS_pannet_multiexm,...
% avg_SCC_pannet_multiexm, std_SCC_pannet_multiexm];
%
%% dicnn: average Q_avg
%avg_Q_dicnn_multiexm = mean(Q_dicnn_multiexm);
%std_Q_dicnn_multiexm = std(Q_dicnn_multiexm);
%
%avg_Q_avg_dicnn_multiexm = mean(Q_avg_dicnn_multiexm);
%std_Q_avg_dicnn_multiexm = std(Q_avg_dicnn_multiexm);
%
%avg_SAM_dicnn_multiexm = mean(SAM_dicnn_multiexm);
%std_SAM_dicnn_multiexm = std(SAM_dicnn_multiexm);
%
%avg_ERGAS_dicnn_multiexm = mean(ERGAS_dicnn_multiexm);
%std_ERGAS_dicnn_multiexm = std(ERGAS_dicnn_multiexm);
%
%avg_SCC_dicnn_multiexm = mean(SCC_dicnn_multiexm);
%std_SCC_dicnn_multiexm = std(SCC_dicnn_multiexm);
%
%Avg_MatrixResults(9,:) = [avg_Q_dicnn_multiexm, std_Q_dicnn_multiexm, avg_Q_avg_dicnn_multiexm, std_Q_avg_dicnn_multiexm, ...
% avg_SAM_dicnn_multiexm, std_SAM_dicnn_multiexm, avg_ERGAS_dicnn_multiexm, std_ERGAS_dicnn_multiexm,...
% avg_SCC_dicnn_multiexm, std_SCC_dicnn_multiexm];
%
%
%
%%
%% msdcnn: average Q_avg
%avg_Q_msdcnn_multiexm = mean(Q_msdcnn_multiexm);
%std_Q_msdcnn_multiexm = std(Q_msdcnn_multiexm);
%
%avg_Q_avg_msdcnn_multiexm = mean(Q_avg_msdcnn_multiexm);
%std_Q_avg_msdcnn_multiexm = std(Q_avg_msdcnn_multiexm);
%
%avg_SAM_msdcnn_multiexm = mean(SAM_msdcnn_multiexm);
%std_SAM_msdcnn_multiexm = std(SAM_msdcnn_multiexm);
%
%avg_ERGAS_msdcnn_multiexm = mean(ERGAS_msdcnn_multiexm);
%std_ERGAS_msdcnn_multiexm = std(ERGAS_msdcnn_multiexm);
%
%avg_SCC_msdcnn_multiexm = mean(SCC_msdcnn_multiexm);
%std_SCC_msdcnn_multiexm = std(SCC_msdcnn_multiexm);
%
%Avg_MatrixResults(10,:) = [avg_Q_msdcnn_multiexm, std_Q_msdcnn_multiexm, avg_Q_avg_msdcnn_multiexm, std_Q_avg_msdcnn_multiexm, ...
% avg_SAM_msdcnn_multiexm, std_SAM_msdcnn_multiexm, avg_ERGAS_msdcnn_multiexm, std_ERGAS_msdcnn_multiexm,...
% avg_SCC_msdcnn_multiexm, std_SCC_msdcnn_multiexm];
%
%
%% bdpn: average Q_avg
%avg_Q_bdpn_multiexm = mean(Q_bdpn_multiexm);
%std_Q_bdpn_multiexm = std(Q_bdpn_multiexm);
%
%avg_Q_avg_bdpn_multiexm = mean(Q_avg_bdpn_multiexm);
%std_Q_avg_bdpn_multiexm = std(Q_avg_bdpn_multiexm);
%
%avg_SAM_bdpn_multiexm = mean(SAM_bdpn_multiexm);
%std_SAM_bdpn_multiexm = std(SAM_bdpn_multiexm);
%
%avg_ERGAS_bdpn_multiexm = mean(ERGAS_bdpn_multiexm);
%std_ERGAS_bdpn_multiexm = std(ERGAS_bdpn_multiexm);
%
%avg_SCC_bdpn_multiexm = mean(SCC_bdpn_multiexm);
%std_SCC_bdpn_multiexm = std(SCC_bdpn_multiexm);
%
%Avg_MatrixResults(11,:) = [avg_Q_bdpn_multiexm, std_Q_bdpn_multiexm, avg_Q_avg_bdpn_multiexm, std_Q_avg_bdpn_multiexm, ...
% avg_SAM_bdpn_multiexm, std_SAM_bdpn_multiexm, avg_ERGAS_bdpn_multiexm, std_ERGAS_bdpn_multiexm,...
% avg_SCC_bdpn_multiexm, std_SCC_bdpn_multiexm];
%
%
%
%% fusionnet: average Q_avg
%avg_Q_fusionnet_multiexm = mean(Q_fusionnet_multiexm);
%std_Q_fusionnet_multiexm = std(Q_fusionnet_multiexm);
%
%avg_Q_avg_fusionnet_multiexm = mean(Q_avg_fusionnet_multiexm);
%std_Q_avg_fusionnet_multiexm = std(Q_avg_fusionnet_multiexm);
%
%avg_SAM_fusionnet_multiexm = mean(SAM_fusionnet_multiexm);
%std_SAM_fusionnet_multiexm = std(SAM_fusionnet_multiexm);
%
%avg_ERGAS_fusionnet_multiexm = mean(ERGAS_fusionnet_multiexm);
%std_ERGAS_fusionnet_multiexm = std(ERGAS_fusionnet_multiexm);
%
%avg_SCC_fusionnet_multiexm = mean(SCC_fusionnet_multiexm);
%std_SCC_fusionnet_multiexm = std(SCC_fusionnet_multiexm);
%
%Avg_MatrixResults(12,:) = [avg_Q_fusionnet_multiexm, std_Q_fusionnet_multiexm, avg_Q_avg_fusionnet_multiexm, std_Q_avg_fusionnet_multiexm, ...
% avg_SAM_fusionnet_multiexm, std_SAM_fusionnet_multiexm, avg_ERGAS_fusionnet_multiexm, std_ERGAS_fusionnet_multiexm,...
% avg_SCC_fusionnet_multiexm, std_SCC_fusionnet_multiexm];
%% lagnet: average Q_avg
%avg_Q_lagnet_multiexm = mean(Q_lagnet_multiexm);
%std_Q_lagnet_multiexm = std(Q_lagnet_multiexm);
%
%avg_Q_avg_lagnet_multiexm = mean(Q_avg_lagnet_multiexm);
%std_Q_avg_lagnet_multiexm = std(Q_avg_lagnet_multiexm);
%
%avg_SAM_lagnet_multiexm = mean(SAM_lagnet_multiexm);
%std_SAM_lagnet_multiexm = std(SAM_lagnet_multiexm);
%
%avg_ERGAS_lagnet_multiexm = mean(ERGAS_lagnet_multiexm);
%std_ERGAS_lagnet_multiexm = std(ERGAS_lagnet_multiexm);
%
%avg_SCC_lagnet_multiexm = mean(SCC_lagnet_multiexm);
%std_SCC_lagnet_multiexm = std(SCC_lagnet_multiexm);
%
%Avg_MatrixResults(13,:) = [avg_Q_lagnet_multiexm, std_Q_lagnet_multiexm, avg_Q_avg_lagnet_multiexm, std_Q_avg_lagnet_multiexm, ...
% avg_SAM_lagnet_multiexm, std_SAM_lagnet_multiexm, avg_ERGAS_lagnet_multiexm, std_ERGAS_lagnet_multiexm,...
% avg_SCC_lagnet_multiexm, std_SCC_lagnet_multiexm];
%matrix2latex(Avg_MatrixResults(:,[1,2, 5,6, 7,8 ]),'Avg_RR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'Q2n'}, {'Q2n-std'}, {'SAM'}, {'SAM-std'}, {'ERGAS'}, {'ERGAS-std'}],'alignment','c','format', '%.4f');
fprintf('\n')
disp('#######################################################')
%disp(['Display the performance for:', num2str(1:i)])
disp('#######################################################')
disp(' |====Q====|===Q_avg===|=====SAM=====|======ERGAS=======|=======SCC=======')
Avg_MatrixResults
|
294coder/Efficient-MIF | 1,176 | Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_unref_batched_images.m | function res = analysis_unref_batched_images(path, ratio, sensor)
if strcmp(sensor, 'QB') || strcmp(sensor, 'GF2')
sensor = 'IKONOS';
end
disp(sensor)
data=load(path);
sr = data.sr;
ms = data.ms;
lms = data.lms;
pan = data.pan;
addpath('./Quality_Indices/')
addpath('./Tools')
sz = size(sr);
bs = sz(1);
d_lambdas = [];
qnr_indices = [];
d_ses = [];
blockSize = 32;
res = {};
h = size(pan, 3);
w = size(pan, 4);
for i = (1: bs)
sr1 = permute(squeeze(sr(i, :, :, :)), [2, 3, 1]);
ms1 = permute(squeeze(ms(i, :, :, :)), [2, 3, 1]);
lms1 = permute(squeeze(lms(i, :, :, :)), [2, 3, 1]);
pan1 = permute(reshape(squeeze(pan(i, :, :, :)), 1, h, w), [2, 3, 1]);
[d_lambdas(i), d_ses(i), qnr_indices(i)] = indexes_evaluation_FS(sr1, ms1, pan1, 11, 0, lms1, sensor, ratio, 0);
fprintf('sample %d - d_lambda: %f, qnr_index: %f, d_s: %f \n', i, d_lambdas(i), qnr_indices(i), d_ses(i))
end
res.d_lambda = [mean(d_lambdas), std(d_lambdas)];
res.qnr_index = [mean(qnr_indices), std(qnr_indices)];
res.d_s = [mean(d_ses), std(d_ses)];
end |
294coder/Efficient-MIF | 1,748 | Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_unref_unbatched_images.m | function res = analysis_unref_unbatched_images(dir_path, ratio, sensor)
full_data_path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_wv3_OrigScale_multiExm1.h5";
sensor = lower(sensor);
% if sensor == "wv3" || sensor == "wv2" || sensor == "qb"
% const = 2047;
% elseif sensor == "gf2" || sensor == "gf"
% const = 1023;
% else
% error(strcat(sensor, ' is not supported!'))
ms_s = h5read(full_data_path, '/ms');
ms_s = permute(ms_s, [4,2,1,3]);
lms_s = h5read(full_data_path, '/lms');
lms_s = permute(lms_s, [4,2,1,3]);
pan_s = h5read(full_data_path, '/pan');
pan_s = permute(pan_s, [4,2,1,3]);
bs = size(lms_s, 1);
addpath('./Tools')
addpath('./Quality_Indices/')
d_lambdas = [];
qnr_indices = [];
d_ses = [];
res = {};
h = size(pan_s, 3);
w = size(pan_s, 4);
for i = (0:bs-1)
p = strcat(dir_path, "/", "output_mulExm_", string(i), ".mat");
sr2 = load(p);
sr = sr2.sr;
i = i+1;
ms = squeeze(ms_s(i, :, :, :));
lms = squeeze(lms_s(i, :, :, :));
pan = squeeze(pan_s(i, :, :, :));
[d_lambdas(i), d_ses(i), qnr_indices(i)] = indexes_evaluation_FS(sr, ms, pan, 11, 0, lms, sensor, ratio, 0);
fprintf('sample %d - d_lambda: %f, qnr_index: %f, d_s: %f \n', i, d_lambdas(i), qnr_indices(i), d_ses(i))
res.d_lambda = [mean(d_lambdas), std(d_lambdas)];
res.qnr_index = [mean(qnr_indices), std(qnr_indices)];
res.d_s = [mean(d_ses), std(d_ses)];
end
res.d_lambda = [mean(d_lambdas), std(d_lambdas)];
res.qnr_index = [mean(qnr_indices), std(qnr_indices)];
res.d_s = [mean(d_ses), std(d_ses)];
end
|
2977094657/DidaAPI | 1,142 | frontend/docs/api/habits.md | # 获取所有习惯
获取当前用户的所有习惯列表。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/habits`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
无需参数,使用当前认证会话。
## 响应格式
### 成功响应
```json
[
{
"id": "string",
"name": "string",
"iconRes": "string",
"color": "string",
"status": 1,
"encouragement": "string",
"totalCheckIns": 0,
"createdTime": "string",
"modifiedTime": "string",
"type": "string",
"goal": 0.0
},
{
"id": "string",
"name": "string",
"iconRes": "string",
"color": "string",
"status": 0,
"encouragement": "string",
"totalCheckIns": 0,
"createdTime": "string",
"modifiedTime": "string",
"type": "string",
"goal": 0.0
}
]
```
## 响应字段说明
| 字段 | 类型 | 说明 |
|------|------|------|
| id | string | 习惯ID |
| name | string | 习惯名称 |
| iconRes | string | 图标资源名称 |
| color | string | 习惯颜色(十六进制) |
| status | number | 状态(0=未完成,1=已完成) |
| encouragement | string | 激励语句 |
| totalCheckIns | number | 总打卡次数 |
| createdTime | string | 创建时间 |
| modifiedTime | string | 修改时间 |
| type | string | 习惯类型(daily/weekly等) |
| goal | number | 目标值 |
|
2977094657/DidaAPI | 2,293 | frontend/docs/api/url-management.md | # URL和外部链接管理
本项目采用统一的URL管理策略,所有外部链接和API端点都在 `core/urls.py` 文件中集中管理。
## 管理原则
### 为什么需要统一管理
1. **避免硬编码** - 防止URL散布在各个文件中
2. **便于维护** - 统一修改和更新链接
3. **提高可读性** - 清晰的分类和注释
4. **便于测试** - 集中的URL便于健康检查
### 分类管理
所有URL按功能分为以下几类:
#### 微信开放平台相关
- `WECHAT_URLS` - 微信API的基础URL
- `WECHAT_CONFIG` - 微信应用配置参数
#### 滴答清单API相关
- `DIDA_API_BASE` - 滴答清单API基础配置
- `DIDA_AUTH_APIS` - 认证相关API端点
- `DIDA_TASK_APIS` - 任务管理API端点
- `DIDA_PROJECT_APIS` - 项目管理API端点
#### 外部文档链接
- `OFFICIAL_DOCS` - 官方文档链接
- `TECH_REFERENCES` - 技术参考文档
## 使用方法
### 导入URL模块
```python
from core import urls
```
### 使用预定义的URL
```python
# 获取微信二维码基础URL
qr_base_url = urls.WECHAT_URLS["qr_base_url"]
# 获取滴答清单API基础URL
api_base = urls.DIDA_API_BASE["base_url"]
```
### 使用URL构建函数
```python
# 构建微信登录二维码URL
qr_url = urls.build_wechat_qr_url(state="Lw==")
# 构建滴答清单API完整URL
api_url = urls.build_dida_api_url("/batch/check/0")
# 构建微信登录验证URL
validate_url = urls.build_wechat_validate_url(code="xxx", state="Lw==")
```
## URL分类详情
### 微信相关URL
```python
WECHAT_URLS = {
"qr_base_url": "https://open.weixin.qq.com/connect/qrconnect",
"qr_image_base_url": "https://open.weixin.qq.com/connect/qrcode",
"poll_login_url": "https://long.open.weixin.qq.com/connect/l/qrconnect",
"redirect_uri": "https://dida365.com/sign/wechat"
}
```
### 滴答清单API端点
```python
DIDA_TASK_APIS = {
"get_all_tasks": "/batch/check/0",
"task_crud": "/task",
"task_search": "/task/search"
}
```
### 官方文档链接
```python
OFFICIAL_DOCS = {
"wechat_login_guide": "https://developers.weixin.qq.com/doc/oplatform/Website_App/WeChat_Login/Wechat_Login",
"dida_official": "https://dida365.com"
}
```
## 辅助函数
### URL构建函数
- `build_wechat_qr_url(state)` - 构建微信登录二维码URL
- `build_wechat_poll_url(uuid, timestamp)` - 构建轮询URL
- `build_dida_api_url(endpoint)` - 构建滴答清单API URL
- `build_wechat_validate_url(code, state)` - 构建验证URL
### 管理函数
- `get_all_external_urls()` - 获取所有外部URL
- `get_api_endpoints()` - 获取所有API端点
## 添加新URL的步骤
### 1. 确定分类
根据URL的用途选择合适的分类字典。
### 2. 添加URL
```python
# 在对应的字典中添加新URL
DIDA_TASK_APIS = {
# 现有URL...
"new_endpoint": "/new/api/endpoint" # 新增
}
```
### 3. 添加注释
为新URL添加清晰的注释说明其用途。
### 4. 更新文档
在相关的API文档中引用新的URL配置。
## 相关文件
- `core/urls.py` - URL管理主文件
- `core/config.py` - 非URL配置管理
- `config.toml` - 配置文件(已移除URL配置)
|
294coder/Efficient-MIF | 23,781 | Pansharpening_Hyper_SR_Matlab_Test_Package/Demo2_Full_Resolution_WV3_multi.m | % LJ Deng(UESTC)
% 2020-06-02
clear; close all;
%% =======load directors========
% Tools
addpath([pwd,'/Tools']);
% L, locatio
% % Select algorithms to run
algorithms = {'BT-H'};%'PNN'
data_name = '3_EPS/WV3/wv3_os_'; % director to save EPS figures
%% ==========Read each Data====================
%% read each data
file_test = '/Data2/ZiHanCao/datasets/pansharpening/qb/full_examples/test_qb_OrigScale_multiExm1.h5';
disp(file_test)
%load(file_test) % get I_MS_LR, I_MS, I_PAN and sensors' info.
ms_multiExm_tmp = h5read(file_test,'/ms'); % WxHxCxN=1x2x3x4
ms_multiExm = permute(ms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
lms_multiExm_tmp = h5read(file_test,'/lms'); % WxHxCxN=1x2x3x4
lms_multiExm = permute(lms_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
pan_multiExm_tmp = h5read(file_test,'/pan'); % WxHxCxN=1x2x3x4
pan_multiExm = permute(pan_multiExm_tmp, [4 2 1 3]); % NxHxWxC=4x2x1x3
%% ==========Read each Data====================
exm_num = size(ms_multiExm, 1);
for i = (1: exm_num)
disp(i)
%% read each data
LRMS_tmp = ms_multiExm(i, :, :, :); % I_MS_LR
I_MS_LR = squeeze(LRMS_tmp);
LMS_tmp = lms_multiExm(i, :, :, :); % I_MS
I_MS = squeeze(LMS_tmp);
PAN_tmp = pan_multiExm(i, :, :, :); % I_PAN
I_PAN = squeeze(PAN_tmp);
NumIndexes = 3;
MatrixResults = zeros(numel(algorithms),NumIndexes);
alg = 0;
flagQNR = 0; %% Flag QNR/HQNR, 1: QNR otherwise HQNR
% for img show
location1 = [10 50 190 240]; %default: data6: [10 50 1 60]; data7:[140 180 5 60]
location2 = [230 280 240 290]; %default: data6: [190 240 5 60]; data7:[190 235 120 150]
sensor = 'GF2';
% disp(sensor)
%% load Indexes for WV3_FR
% sensor = 'WV3';
Qblocks_size = 32;
bicubic = 0;% Interpolator
flag_cut_bounds = 1;% Cut Final Image
dim_cut = 21;% Cut Final Image
thvalues = 0;% Threshold values out of dynamic range
printEPS = 0;% Print Eps
ratio = 4;% Resize Factor
L = 11;% Radiometric Resolution
clear print
%% show I_MS_LR, I_GT, PAN Imgs:
% if size(I_MS,3) == 4
% showImage4LR(I_MS_LR,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% else
% showImage8_zoomin(I_MS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% end
%
% %showPan(I_PAN,printEPS,2,flag_cut_bounds,dim_cut);
% showPan_zoomin(I_PAN,printEPS,2,flag_cut_bounds,dim_cut, location1, location2);
% pause(2);print('-deps', strcat(data_name, num2str(i-1), '_pan', '.eps'))
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%% CS-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%%
%% ====== 1) BT-H Method ======
if ismember('BT-H',algorithms)
alg = alg + 1;
cd BT-H
t2=tic;
% I_BT_H = BroveyRegHazeMin(I_MS,I_PAN,ratio);
% time_BT_H = toc(t2);
% fprintf('Elaboration time BT-H: %.2f [sec]\n',time_BT_H);
cd ..
%%% Quality indexes computation
[D_lambda_BT_H, D_S_BT_H, QNRI_BT_H] = indexes_evaluation_FS(I_MS,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
MatrixResults(alg,:) = [D_lambda_BT_H,D_S_BT_H,QNRI_BT_H];
% MatrixImage(:,:,:,alg) = I_BT_H;
D_lambda_BT_H_multiexm(i) = D_lambda_BT_H;
D_S_BT_H_multiexm(i) = D_S_BT_H;
QNRI_BT_H_multiexm(i) = QNRI_BT_H;
%
% showImage8_zoomin(I_BT_H,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bth.eps'))
end
%% ====== 2) BDSD-PC Method ======
if ismember('BDSD-PC',algorithms)
alg = alg + 1;
cd BDSD
t2=tic;
I_BDSD_PC = BDSD_PC(I_MS,I_PAN,ratio,sensor);
time_BDSD_PC = toc(t2);
fprintf('Elaboration time BDSD-PC: %.2f [sec]\n',time_BDSD_PC);
cd ..
[D_lambda_BDSD_PC, D_S_BDSD_PC, QNRI_BDSD_PC] = indexes_evaluation_FS(I_BDSD_PC,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
MatrixResults(alg,:) = [D_lambda_BDSD_PC,D_S_BDSD_PC,QNRI_BDSD_PC];
MatrixImage(:,:,:,alg) = I_BDSD_PC;
D_lambda_BDSD_PC_multiexm(i) = D_lambda_BDSD_PC;
D_S_BDSD_PC_multiexm(i) = D_S_BDSD_PC;
QNRI_BDSD_PC_multiexm(i) = QNRI_BDSD_PC;
% showImage8_zoomin(I_BDSD_PC,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_bdsd_pc.eps'))
end
% %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% %%%%%%%%%%%%% MRA-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%% %
%% ====== 1) MTF-GLP-HPM-R Method ======
if ismember('MTF-GLP-HPM-R',algorithms)
alg = alg + 1;
cd GLP
t2=tic;
I_MTF_GLP_HPM_R = MTF_GLP_HPM_R(I_MS,I_PAN,sensor,ratio);
time_MTF_GLP_HPM_R = toc(t2);
fprintf('Elaboration time MTF-GLP: %.2f [sec]\n',time_MTF_GLP_HPM_R);
cd ..
[D_lambda_MTF_GLP_HPM_R, D_S_MTF_GLP_HPM_R, QNRI_MTF_GLP_HPM_R] = indexes_evaluation_FS(I_MTF_GLP_HPM_R,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
MatrixResults(alg,:) = [D_lambda_MTF_GLP_HPM_R,D_S_MTF_GLP_HPM_R,QNRI_MTF_GLP_HPM_R];
MatrixImage(:,:,:,alg) = I_MTF_GLP_HPM_R;
D_lambda_MTF_GLP_HPM_R_multiexm(i) = D_lambda_MTF_GLP_HPM_R;
D_S_MTF_GLP_HPM_R_multiexm(i) = D_S_MTF_GLP_HPM_R;
QNRI_MTF_GLP_HPM_R_multiexm(i) = QNRI_MTF_GLP_HPM_R;
% showImage8_zoomin(I_MTF_GLP_HPM_R,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglp_hpm_r.eps'))
end
%% ====== 2) MTF-GLP-FS Method ======
if ismember('MTF-GLP-FS',algorithms)
alg = alg + 1;
cd GLP
t2=tic;
I_MTF_GLP_FS = MTF_GLP_FS(I_MS,I_PAN,sensor,ratio);
time_MTF_GLP_FS = toc(t2);
fprintf('Elaboration time MTF-GLP-FS: %.2f [sec]\n',time_MTF_GLP_FS);
cd ..
[D_lambda_MTF_GLP_FS, D_S_MTF_GLP_FS, QNRI_MTF_GLP_FS] = indexes_evaluation_FS(I_MTF_GLP_FS,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
MatrixResults(alg,:) = [D_lambda_MTF_GLP_FS,D_S_MTF_GLP_FS,QNRI_MTF_GLP_FS];
MatrixImage(:,:,:,alg) = I_MTF_GLP_FS;
D_lambda_MTF_GLP_FS_multiexm(i) = D_lambda_MTF_GLP_FS;
D_S_MTF_GLP_FS_multiexm(i) = D_S_MTF_GLP_FS;
QNRI_MTF_GLP_FS_multiexm(i) = QNRI_MTF_GLP_FS;
% showImage8_zoomin(I_MTF_GLP_FS,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_mtfglpfs.eps'))
end
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%% VO-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%%
%% ====== 1) TV Method ======
if ismember('TV',algorithms)
alg = alg + 1;
%%%%%%%%%%%%%%%%%%%%%%%%%% Parameters setting %%%%%%%%%%%%%%%%%%%%%%%%%%%%%
switch sensor
case 'IKONOS'
w=[0.1091 0.2127 0.2928 0.3854];
c = 8;
alpha=1.064;
maxiter=10;
lambda = 0.47106;
case {'GeoEye1','WV4'}
w=[0.1552, 0.3959, 0.2902, 0.1587];
c = 8;
alpha=0.75;
maxiter=50;
lambda = 157.8954;
case 'WV3'
w=[0.0657 0.1012 0.1537 0.1473 0.1245 0.1545 0.1338 0.1192];
c = 8;
alpha=0.75;
maxiter=50;
lambda = 1.0000e-03;
end
cd TV
t2 = tic;
I_TV = TV_pansharpen(I_MS_LR,I_PAN,alpha,lambda,c,maxiter,w);
time_TV = toc(t2);
fprintf('Elaboration time TV: %.2f [sec]\n',time_TV);
cd ..
[D_lambda_TV, D_S_TV, QNRI_TV] = indexes_evaluation_FS(I_TV,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
MatrixResults(alg,:) = [D_lambda_TV,D_S_TV,QNRI_TV];
MatrixImage(:,:,:,alg) = I_TV;
D_lambda_TV_multiexm(i) = D_lambda_TV;
D_S_TV_multiexm(i) = D_S_TV;
QNRI_TV_multiexm(i) = QNRI_TV;
%
% showImage8_zoomin(I_TV,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1), '_tv.eps'))
end
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%% DL-based Methods %%%%%%%%%%%%%%%%%%%%%%%%%%
% %% ====== 1) PNN Method ======
% if ismember('PNN',algorithms)
%
% load(strcat('2_DL_Result/WV3_Full/PNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PNN
% I_pnn = double(sr);
%
%
% alg = alg + 1;
% [D_lambda_pnn, D_S_pnn, QNRI_pnn] = indexes_evaluation_FS(I_pnn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
% MatrixResults(alg,:) = [D_lambda_pnn,D_S_pnn,QNRI_pnn];
% MatrixImage(:,:,:,alg) = I_pnn;
%
% D_lambda_pnn_multiexm(i) = D_lambda_pnn;
% D_S_pnn_multiexm(i) = D_S_pnn;
% QNRI_pnn_multiexm(i) = QNRI_pnn;
%%
%% showImage8_zoomin(I_pnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
%% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pnn.eps'))
% end
%
% %% ====== 2) PanNet Method ======
%
% if ismember('PanNet',algorithms)
% load(strcat('2_DL_Result/WV3_Full/PanNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for PanNet
% I_pannet = double(sr);
%
% alg = alg + 1;
% [D_lambda_pannet, D_S_pannet, QNRI_pannet] = indexes_evaluation_FS(I_pannet,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
% MatrixResults(alg,:) = [D_lambda_pannet,D_S_pannet,QNRI_pannet];
% MatrixImage(:,:,:,alg) = I_pannet;
%
%
% D_lambda_pannet_multiexm(i) = D_lambda_pannet;
% D_S_pannet_multiexm(i) = D_S_pannet;
% QNRI_pannet_multiexm(i) = QNRI_pannet;
%%
% showImage8_zoomin(I_pannet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_pannet.eps'))
% end
%
%
%
%%% ====== 3) DiCNN Method ======
% if ismember('DiCNN',algorithms)
% load(strcat('2_DL_Result/WV3_Full/DiCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for DiCNN
% I_dicnn = double(sr);
%
%
% alg = alg + 1;
% [D_lambda_dicnn, D_S_dicnn, QNRI_dicnn] = indexes_evaluation_FS(I_dicnn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
% MatrixResults(alg,:) = [D_lambda_dicnn,D_S_dicnn,QNRI_dicnn];
% MatrixImage(:,:,:,alg) = I_dicnn;
%
% D_lambda_dicnn_multiexm(i) = D_lambda_dicnn;
% D_S_dicnn_multiexm(i) = D_S_dicnn;
% QNRI_dicnn_multiexm(i) = QNRI_dicnn;
%%
% showImage8_zoomin(I_dicnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_dicnn.eps'))
% end
% %% ====== 4) MSDCNN Method ======
%
%if ismember('MSDCNN',algorithms)
% load(strcat('2_DL_Result/WV3_Full/MSDCNN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for MSDCNN
% I_msdcnn = double(sr);
%
%
% alg = alg + 1;
% [D_lambda_msdcnn, D_S_msdcnn, QNRI_msdcnn] = indexes_evaluation_FS(I_msdcnn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
% MatrixResults(alg,:) = [D_lambda_msdcnn,D_S_msdcnn,QNRI_msdcnn];
% MatrixImage(:,:,:,alg) = I_msdcnn;
%
% D_lambda_msdcnn_multiexm(i) = D_lambda_msdcnn;
% D_S_msdcnn_multiexm(i) = D_S_msdcnn;
% QNRI_msdcnn_multiexm(i) = QNRI_msdcnn;
% showImage8_zoomin(I_msdcnn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_msdcnn.eps'))
%end
% %% ====== 5) BDPN Method ======
%
%
% if ismember('BDPN',algorithms)
% load(strcat('2_DL_Result/WV3_Full/BDPN/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for BDPN
% I_bdpn = double(sr);
%
%
% alg = alg + 1;
% [D_lambda_bdpn, D_S_bdpn, QNRI_bdpn] = indexes_evaluation_FS(I_bdpn,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
% MatrixResults(alg,:) = [D_lambda_bdpn,D_S_bdpn,QNRI_bdpn];
% MatrixImage(:,:,:,alg) = I_bdpn;
%
% D_lambda_bdpn_multiexm(i) = D_lambda_bdpn;
% D_S_bdpn_multiexm(i) = D_S_bdpn;
% QNRI_bdpn_multiexm(i) = QNRI_bdpn;
%
%
% showImage8_zoomin(I_bdpn,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_bdpn.eps'))
% end
% %% ====== 6) FusionNet Method ======
% if ismember('FusionNet',algorithms)
% file_fusionnet = 'fusionnet_wv3_rs';
% load(strcat('2_DL_Result/WV3_Full/FusionNet/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for FusionNet
% I_fusionnet = double(sr);
%
%
% alg = alg + 1;
% [D_lambda_fusionnet, D_S_fusionnet, QNRI_fusionnet] = indexes_evaluation_FS(I_fusionnet,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
% MatrixResults(alg,:) = [D_lambda_fusionnet,D_S_fusionnet,QNRI_fusionnet];
% MatrixImage(:,:,:,alg) = I_fusionnet;
%
% D_lambda_fusionnet_multiexm(i) = D_lambda_fusionnet;
% D_S_fusionnet_multiexm(i) = D_S_fusionnet;
% QNRI_fusionnet_multiexm(i) = QNRI_fusionnet;
%
% showImage8_zoomin(I_fusionnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_fusionnet.eps'))
% end
% %% ====== 7) LAGConv Method ======
%
% if ismember('LAGConv',algorithms)
% load(strcat('2_DL_Result/WV3_Full/LAGConv/results/output_mulExm_', num2str(i-1), '.mat')) % load i-th image for LAGConv
% I_lagnet = double(sr);
% alg = alg + 1;
% [D_lambda_lagnet, D_S_lagnet, QNRI_lagnet] = indexes_evaluation_FS(I_lagnet,I_MS_LR,I_PAN,L,thvalues,I_MS,sensor,ratio,flagQNR);
% MatrixResults(alg,:) = [D_lambda_lagnet,D_S_lagnet,QNRI_lagnet];
% MatrixImage(:,:,:,alg) = I_lagnet;
%
% D_lambda_lagnet_multiexm(i) = D_lambda_lagnet;
% D_S_lagnet_multiexm(i) = D_S_lagnet;
% QNRI_lagnet_multiexm(i) = QNRI_lagnet;
%%
% showImage8_zoomin(I_lagnet,printEPS,1,flag_cut_bounds,dim_cut,thvalues,L, location1, location2);
% pause(2);print('-depsc', strcat(data_name, num2str(i-1),'_lagnet.eps'))
% end
end
%% Print in LATEX
% if flagQNR == 1
% matrix2latex(MatrixResults,'FR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'DL'},{'DS'},{'QNR'}],'alignment','c','format', '%.4f');
% else
% matrix2latex(MatrixResults,'FR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'DL'},{'DS'},{'HQNR'}],'alignment','c','format', '%.4f');
% end
%% View All
if size(I_MS,3) == 4
vect_index_RGB = [3,2,1];
else
vect_index_RGB = [5,3,2];
end
titleImages = algorithms;
% figure, showImagesAll(MatrixImage,titleImages,vect_index_RGB,flag_cut_bounds,dim_cut,0);
% BT_H: average D_lambda_avg
avg_D_lambda_BT_H_multiexm = mean(D_lambda_BT_H_multiexm);
std_D_lambda_BT_H_multiexm = std(D_lambda_BT_H_multiexm);
avg_D_S_BT_H_multiexm = mean(D_S_BT_H_multiexm);
std_D_S_BT_H_multiexm = std(D_S_BT_H_multiexm);
avg_QNRI_BT_H_multiexm = mean(QNRI_BT_H_multiexm);
std_QNRI_BT_H_multiexm = std(QNRI_BT_H_multiexm);
Avg_MatrixResults(1,:) = [avg_D_lambda_BT_H_multiexm, std_D_lambda_BT_H_multiexm, ...
avg_D_S_BT_H_multiexm, std_D_S_BT_H_multiexm, avg_QNRI_BT_H_multiexm, std_QNRI_BT_H_multiexm
];
% BDSD: average D_lambda_avg
% avg_D_lambda_BDSD_PC_multiexm = mean(D_lambda_BDSD_PC_multiexm);
% std_D_lambda_BDSD_PC_multiexm = std(D_lambda_BDSD_PC_multiexm);
% avg_D_S_BDSD_PC_multiexm = mean(D_S_BDSD_PC_multiexm);
% std_D_S_BDSD_PC_multiexm = std(D_S_BDSD_PC_multiexm);
% avg_QNRI_BDSD_PC_multiexm = mean(QNRI_BDSD_PC_multiexm);
% std_QNRI_BDSD_PC_multiexm = std(QNRI_BDSD_PC_multiexm);
% Avg_MatrixResults(2,:) = [avg_D_lambda_BDSD_PC_multiexm, std_D_lambda_BDSD_PC_multiexm, ...
% avg_D_S_BDSD_PC_multiexm, std_D_S_BDSD_PC_multiexm, avg_QNRI_BDSD_PC_multiexm, std_QNRI_BDSD_PC_multiexm
% ];
% % MTF_GLP: average D_lambda_avg
% avg_D_lambda_MTF_GLP_HPM_R_multiexm = mean(D_lambda_MTF_GLP_HPM_R_multiexm);
% std_D_lambda_MTF_GLP_HPM_R_multiexm = std(D_lambda_MTF_GLP_HPM_R_multiexm);
% avg_D_S_MTF_GLP_HPM_R_multiexm = mean(D_S_MTF_GLP_HPM_R_multiexm);
% std_D_S_MTF_GLP_HPM_R_multiexm = std(D_S_MTF_GLP_HPM_R_multiexm);
% avg_QNRI_MTF_GLP_HPM_R_multiexm = mean(QNRI_MTF_GLP_HPM_R_multiexm);
% std_QNRI_MTF_GLP_HPM_R_multiexm = std(QNRI_MTF_GLP_HPM_R_multiexm);
% Avg_MatrixResults(3,:) = [avg_D_lambda_MTF_GLP_HPM_R_multiexm, std_D_lambda_MTF_GLP_HPM_R_multiexm, ...
% avg_D_S_MTF_GLP_HPM_R_multiexm, std_D_S_MTF_GLP_HPM_R_multiexm, avg_QNRI_MTF_GLP_HPM_R_multiexm, std_QNRI_MTF_GLP_HPM_R_multiexm
% ];
% % MTF_GLP_FS: average D_lambda_avg
% avg_D_lambda_MTF_GLP_FS_multiexm = mean(D_lambda_MTF_GLP_FS_multiexm);
% std_D_lambda_MTF_GLP_FS_multiexm = std(D_lambda_MTF_GLP_FS_multiexm);
% avg_D_S_MTF_GLP_FS_multiexm = mean(D_S_MTF_GLP_FS_multiexm);
% std_D_S_MTF_GLP_FS_multiexm = std(D_S_MTF_GLP_FS_multiexm);
% avg_QNRI_MTF_GLP_FS_multiexm = mean(QNRI_MTF_GLP_FS_multiexm);
% std_QNRI_MTF_GLP_FS_multiexm = std(QNRI_MTF_GLP_FS_multiexm);
% Avg_MatrixResults(4,:) = [avg_D_lambda_MTF_GLP_FS_multiexm, std_D_lambda_MTF_GLP_FS_multiexm, ...
% avg_D_S_MTF_GLP_FS_multiexm, std_D_S_MTF_GLP_FS_multiexm, avg_QNRI_MTF_GLP_FS_multiexm, std_QNRI_MTF_GLP_FS_multiexm
% ];
% % TV: average D_lambda_avg
% avg_D_lambda_TV_multiexm = mean(D_lambda_TV_multiexm);
% std_D_lambda_TV_multiexm = std(D_lambda_TV_multiexm);
% avg_D_S_TV_multiexm = mean(D_S_TV_multiexm);
% std_D_S_TV_multiexm = std(D_S_TV_multiexm);
% avg_QNRI_TV_multiexm = mean(QNRI_TV_multiexm);
% std_QNRI_TV_multiexm = std(QNRI_TV_multiexm);
% Avg_MatrixResults(5,:) = [avg_D_lambda_TV_multiexm, std_D_lambda_TV_multiexm, ...
% avg_D_S_TV_multiexm, std_D_S_TV_multiexm, avg_QNRI_TV_multiexm, std_QNRI_TV_multiexm];
%% pnn: average D_lambda_avg
%avg_D_lambda_pnn_multiexm = mean(D_lambda_pnn_multiexm);
%std_D_lambda_pnn_multiexm = std(D_lambda_pnn_multiexm);
%
%avg_D_S_pnn_multiexm = mean(D_S_pnn_multiexm);
%std_D_S_pnn_multiexm = std(D_S_pnn_multiexm);
%
%avg_QNRI_pnn_multiexm = mean(QNRI_pnn_multiexm);
%std_QNRI_pnn_multiexm = std(QNRI_pnn_multiexm);
%
%
%Avg_MatrixResults(6,:) = [avg_D_lambda_pnn_multiexm, std_D_lambda_pnn_multiexm, ...
% avg_D_S_pnn_multiexm, std_D_S_pnn_multiexm, avg_QNRI_pnn_multiexm, std_QNRI_pnn_multiexm
% ];
%% pannet: average D_lambda_avg
%avg_D_lambda_pannet_multiexm = mean(D_lambda_pannet_multiexm);
%std_D_lambda_pannet_multiexm = std(D_lambda_pannet_multiexm);
%
%
%avg_D_S_pannet_multiexm = mean(D_S_pannet_multiexm);
%std_D_S_pannet_multiexm = std(D_S_pannet_multiexm);
%
%avg_QNRI_pannet_multiexm = mean(QNRI_pannet_multiexm);
%std_QNRI_pannet_multiexm = std(QNRI_pannet_multiexm);
%
%
%Avg_MatrixResults(7,:) = [avg_D_lambda_pannet_multiexm, std_D_lambda_pannet_multiexm, ...
% avg_D_S_pannet_multiexm, std_D_S_pannet_multiexm, avg_QNRI_pannet_multiexm, std_QNRI_pannet_multiexm
% ];
%
%
%
%% dicnn: average D_lambda_avg
%avg_D_lambda_dicnn_multiexm = mean(D_lambda_dicnn_multiexm);
%std_D_lambda_dicnn_multiexm = std(D_lambda_dicnn_multiexm);
%
%avg_D_S_dicnn_multiexm = mean(D_S_dicnn_multiexm);
%std_D_S_dicnn_multiexm = std(D_S_dicnn_multiexm);
%
%avg_QNRI_dicnn_multiexm = mean(QNRI_dicnn_multiexm);
%std_QNRI_dicnn_multiexm = std(QNRI_dicnn_multiexm);
%
%Avg_MatrixResults(8,:) = [avg_D_lambda_dicnn_multiexm, std_D_lambda_dicnn_multiexm, ...
% avg_D_S_dicnn_multiexm, std_D_S_dicnn_multiexm, avg_QNRI_dicnn_multiexm, std_QNRI_dicnn_multiexm
% ];
%% msdcnn: average D_lambda_avg
%avg_D_lambda_msdcnn_multiexm = mean(D_lambda_msdcnn_multiexm);
%std_D_lambda_msdcnn_multiexm = std(D_lambda_msdcnn_multiexm);
%
%
%avg_D_S_msdcnn_multiexm = mean(D_S_msdcnn_multiexm);
%std_D_S_msdcnn_multiexm = std(D_S_msdcnn_multiexm);
%
%avg_QNRI_msdcnn_multiexm = mean(QNRI_msdcnn_multiexm);
%std_QNRI_msdcnn_multiexm = std(QNRI_msdcnn_multiexm);
%
%Avg_MatrixResults(9,:) = [avg_D_lambda_msdcnn_multiexm, std_D_lambda_msdcnn_multiexm, ...
% avg_D_S_msdcnn_multiexm, std_D_S_msdcnn_multiexm, avg_QNRI_msdcnn_multiexm, std_QNRI_msdcnn_multiexm
% ];
%% bdpn: average D_lambda_avg
%avg_D_lambda_bdpn_multiexm = mean(D_lambda_bdpn_multiexm);
%std_D_lambda_bdpn_multiexm = std(D_lambda_bdpn_multiexm);
%
%
%avg_D_S_bdpn_multiexm = mean(D_S_bdpn_multiexm);
%std_D_S_bdpn_multiexm = std(D_S_bdpn_multiexm);
%
%avg_QNRI_bdpn_multiexm = mean(QNRI_bdpn_multiexm);
%std_QNRI_bdpn_multiexm = std(QNRI_bdpn_multiexm);
%
%Avg_MatrixResults(10,:) = [avg_D_lambda_bdpn_multiexm, std_D_lambda_bdpn_multiexm, ...
% avg_D_S_bdpn_multiexm, std_D_S_bdpn_multiexm, avg_QNRI_bdpn_multiexm, std_QNRI_bdpn_multiexm
% ];
%
%
%
%
%%
%% fusionnet: average D_lambda_avg
%avg_D_lambda_fusionnet_multiexm = mean(D_lambda_fusionnet_multiexm);
%std_D_lambda_fusionnet_multiexm = std(D_lambda_fusionnet_multiexm);
%
%avg_D_S_fusionnet_multiexm = mean(D_S_fusionnet_multiexm);
%std_D_S_fusionnet_multiexm = std(D_S_fusionnet_multiexm);
%
%avg_QNRI_fusionnet_multiexm = mean(QNRI_fusionnet_multiexm);
%std_QNRI_fusionnet_multiexm = std(QNRI_fusionnet_multiexm);
%
%Avg_MatrixResults(11,:) = [avg_D_lambda_fusionnet_multiexm, std_D_lambda_fusionnet_multiexm, ...
% avg_D_S_fusionnet_multiexm, std_D_S_fusionnet_multiexm, avg_QNRI_fusionnet_multiexm, std_QNRI_fusionnet_multiexm
% ];
%% lagnet: average D_lambda_avg
%avg_D_lambda_lagnet_multiexm = mean(D_lambda_lagnet_multiexm);
%std_D_lambda_lagnet_multiexm = std(D_lambda_lagnet_multiexm);
%
%avg_D_S_lagnet_multiexm = mean(D_S_lagnet_multiexm);
%std_D_S_lagnet_multiexm = std(D_S_lagnet_multiexm);
%
%avg_QNRI_lagnet_multiexm = mean(QNRI_lagnet_multiexm);
%std_QNRI_lagnet_multiexm = std(QNRI_lagnet_multiexm);
%
%
%Avg_MatrixResults(12,:) = [avg_D_lambda_lagnet_multiexm, std_D_lambda_lagnet_multiexm, ...
% avg_D_S_lagnet_multiexm, std_D_S_lagnet_multiexm, avg_QNRI_lagnet_multiexm, std_QNRI_lagnet_multiexm
% ];
matrix2latex(Avg_MatrixResults,'Avg_RR_Assessment.tex', 'rowLabels',algorithms,'columnLabels',[{'D_lambda'}, {'D_l-std'}, {'D_S'}, {'D_S-std'}, {'QNRI'}, {'QNRI-std'}],'alignment','c','format', '%.4f');
fprintf('\n')
disp('#######################################################')
disp(['Display the performance for:', num2str(1:i)])
disp('#######################################################')
disp(' |===D_lambda_avg===|=====D_s=====|======QNR=======')
Avg_MatrixResults
|
281677160/openwrt-package | 86,169 | luci-app-passwall/root/usr/share/passwall/nftables.sh | #!/bin/bash
DIR="$(cd "$(dirname "$0")" && pwd)"
MY_PATH=$DIR/nftables.sh
NFTABLE_NAME="inet passwall"
NFTSET_LOCAL="passwall_local"
NFTSET_LAN="passwall_lan"
NFTSET_VPS="passwall_vps"
NFTSET_SHUNT="passwall_shunt"
NFTSET_GFW="passwall_gfw"
NFTSET_CHN="passwall_chn"
NFTSET_BLACK="passwall_black"
NFTSET_WHITE="passwall_white"
NFTSET_BLOCK="passwall_block"
NFTSET_LOCAL6="passwall_local6"
NFTSET_LAN6="passwall_lan6"
NFTSET_VPS6="passwall_vps6"
NFTSET_SHUNT6="passwall_shunt6"
NFTSET_GFW6="passwall_gfw6"
NFTSET_CHN6="passwall_chn6"
NFTSET_BLACK6="passwall_black6"
NFTSET_WHITE6="passwall_white6"
NFTSET_BLOCK6="passwall_block6"
FORCE_INDEX=0
USE_SHUNT_TCP=0
USE_SHUNT_UDP=0
. /lib/functions/network.sh
FWI=$(uci -q get firewall.passwall.path 2>/dev/null)
FAKE_IP="198.18.0.0/15"
FAKE_IP_6="fc00::/18"
factor() {
local ports="$1"
if [ -z "$1" ] || [ -z "$2" ] || [ "$ports" = "1:65535" ]; then
echo ""
# acl mac address
elif echo "$1" | grep -qE '([A-Fa-f0-9]{2}:){5}[A-Fa-f0-9]{2}'; then
echo "$2 {$1}"
else
ports=$(echo "$ports" | tr -d ' ' | sed 's/:/-/g' | tr ',' '\n' | awk '!a[$0]++' | grep -v '^$')
[ -z "$ports" ] && { echo ""; return; }
if echo "$ports" | grep -q '^1-65535$'; then
echo ""
return
fi
local port
local port_list=""
for port in $ports; do
port_list="${port_list},$port"
done
port_list="${port_list#,}"
echo "$2 {$port_list}"
fi
}
insert_rule_before() {
[ $# -ge 4 ] || {
return 1
}
local table_name="${1}"; shift
local chain_name="${1}"; shift
local keyword="${1}"; shift
local rule="${1}"; shift
local default_index="${1}"; shift
default_index=${default_index:-0}
local _index=$(nft -a list chain $table_name $chain_name 2>/dev/null | grep "$keyword" | awk -F '# handle ' '{print$2}' | head -n 1 | awk '{print $1}')
if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then
nft "add rule $table_name $chain_name $rule"
else
if [ -z "${_index}" ]; then
_index=${default_index}
fi
nft "insert rule $table_name $chain_name position $_index $rule"
fi
}
insert_rule_after() {
[ $# -ge 4 ] || {
return 1
}
local table_name="${1}"; shift
local chain_name="${1}"; shift
local keyword="${1}"; shift
local rule="${1}"; shift
local default_index="${1}"; shift
default_index=${default_index:-0}
local _index=$(nft -a list chain $table_name $chain_name 2>/dev/null | grep "$keyword" | awk -F '# handle ' '{print$2}' | head -n 1 | awk '{print $1}')
if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then
nft "add rule $table_name $chain_name $rule"
else
if [ -n "${_index}" ]; then
_index=$((_index + 1))
else
_index=${default_index}
fi
nft "insert rule $table_name $chain_name position $_index $rule"
fi
}
RULE_LAST_INDEX() {
[ $# -ge 3 ] || {
echolog "索引列举方式不正确(nftables),终止执行!"
return 1
}
local table_name="${1}"; shift
local chain_name="${1}"; shift
local keyword="${1}"; shift
local default="${1:-0}"; shift
local _index=$(nft -a list chain $table_name $chain_name 2>/dev/null | grep "$keyword" | awk -F '# handle ' '{print$2}' | head -n 1 | awk '{print $1}')
echo "${_index:-${default}}"
}
REDIRECT() {
local s="counter redirect"
[ -n "$1" ] && {
local s="$s to :$1"
[ "$2" == "MARK" ] && s="counter meta mark set $1"
[ "$2" == "TPROXY" ] && {
s="counter meta mark 1 tproxy to :$1"
}
[ "$2" == "TPROXY4" ] && {
s="counter meta mark 1 tproxy ip to :$1"
}
[ "$2" == "TPROXY6" ] && {
s="counter meta mark 1 tproxy ip6 to :$1"
}
}
echo $s
}
destroy_nftset() {
for i in "$@"; do
nft flush set $NFTABLE_NAME $i 2>/dev/null
nft delete set $NFTABLE_NAME $i 2>/dev/null
done
}
gen_nft_tables() {
if ! nft list tables | grep -q "^table inet passwall$"; then
nft -f - <<-EOF
table $NFTABLE_NAME {
chain dstnat {
type nat hook prerouting priority dstnat - 1; policy accept;
}
chain mangle_prerouting {
type filter hook prerouting priority mangle - 1; policy accept;
}
chain mangle_output {
type route hook output priority mangle - 1; policy accept;
}
chain nat_output {
type nat hook output priority -1; policy accept;
}
}
EOF
fi
}
insert_nftset() {
local nftset_name="${1}"; shift
local timeout_argument="${1}"; shift
local default_timeout_argument="365d"
[ -n "${1}" ] && {
local nftset_elements
case "$timeout_argument" in
"-1") nftset_elements=$(echo -e $@ | sed 's/\s/, /g') ;;
"0") nftset_elements=$(echo -e $@ | sed "s/\s/ timeout $default_timeout_argument, /g" | sed "s/$/ timeout $default_timeout_argument/") ;;
*) nftset_elements=$(echo -e $@ | sed "s/\s/ timeout $timeout_argument, /g" | sed "s/$/ timeout $timeout_argument/") ;;
esac
mkdir -p $TMP_PATH2/nftset
cat > "$TMP_PATH2/nftset/$nftset_name" <<-EOF
define $nftset_name = {$nftset_elements}
add element $NFTABLE_NAME $nftset_name \$$nftset_name
EOF
nft -f "$TMP_PATH2/nftset/$nftset_name"
rm -rf "$TMP_PATH2/nftset"
}
}
gen_nftset() {
local nftset_name="${1}"; shift
local ip_type="${1}"; shift
# 0 - don't set defalut timeout
local timeout_argument_set="${1}"; shift
# 0 - don't let element timeout(365 days) when set's timeout parameters be seted
# -1 - follow the set's timeout parameters
local timeout_argument_element="${1}"; shift
nft "list set $NFTABLE_NAME $nftset_name" &>/dev/null
if [ $? -ne 0 ]; then
if [ "$timeout_argument_set" == "0" ]; then
nft "add set $NFTABLE_NAME $nftset_name { type $ip_type; flags interval, timeout; auto-merge; }"
else
nft "add set $NFTABLE_NAME $nftset_name { type $ip_type; flags interval, timeout; timeout $timeout_argument_set; gc-interval $timeout_argument_set; auto-merge; }"
fi
fi
[ -n "${1}" ] && insert_nftset $nftset_name $timeout_argument_element $@
}
get_jump_ipt() {
case "$1" in
direct)
echo "mark != 1 counter return"
;;
proxy)
if [ -n "$2" ] && [ -n "$(echo $2 | grep "^counter")" ]; then
echo "$2"
else
echo "$(REDIRECT $2 $3)"
fi
;;
esac
}
gen_lanlist() {
cat $RULES_PATH/lanlist_ipv4 | tr -s '\n' | grep -v "^#"
}
gen_lanlist_6() {
cat $RULES_PATH/lanlist_ipv6 | tr -s '\n' | grep -v "^#"
}
get_wan_ip() {
local NET_IF
local NET_ADDR
network_flush_cache
network_find_wan NET_IF
network_get_ipaddr NET_ADDR "${NET_IF}"
echo $NET_ADDR
}
get_wan6_ip() {
local NET_IF
local NET_ADDR
network_flush_cache
network_find_wan6 NET_IF
network_get_ipaddr6 NET_ADDR "${NET_IF}"
echo $NET_ADDR
}
load_acl() {
([ "$ENABLED_ACLS" == 1 ] || ([ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ])) && echolog " - 访问控制:"
[ "$ENABLED_ACLS" == 1 ] && {
acl_app
for sid in $(ls -F ${TMP_ACL_PATH} | grep '/$' | awk -F '/' '{print $1}' | grep -v 'default'); do
eval $(uci -q show "${CONFIG}.${sid}" | cut -d'.' -sf 3-)
tcp_no_redir_ports=${tcp_no_redir_ports:-default}
udp_no_redir_ports=${udp_no_redir_ports:-default}
use_global_config=${use_global_config:-0}
tcp_proxy_drop_ports=${tcp_proxy_drop_ports:-default}
udp_proxy_drop_ports=${udp_proxy_drop_ports:-default}
tcp_redir_ports=${tcp_redir_ports:-default}
udp_redir_ports=${udp_redir_ports:-default}
use_direct_list=${use_direct_list:-1}
use_proxy_list=${use_proxy_list:-1}
use_block_list=${use_block_list:-1}
use_gfw_list=${use_gfw_list:-1}
chn_list=${chn_list:-direct}
tcp_proxy_mode=${tcp_proxy_mode:-proxy}
udp_proxy_mode=${udp_proxy_mode:-proxy}
[ "$tcp_no_redir_ports" = "default" ] && tcp_no_redir_ports=$TCP_NO_REDIR_PORTS
[ "$udp_no_redir_ports" = "default" ] && udp_no_redir_ports=$UDP_NO_REDIR_PORTS
[ "$tcp_proxy_drop_ports" = "default" ] && tcp_proxy_drop_ports=$TCP_PROXY_DROP_PORTS
[ "$udp_proxy_drop_ports" = "default" ] && udp_proxy_drop_ports=$UDP_PROXY_DROP_PORTS
[ "$tcp_redir_ports" = "default" ] && tcp_redir_ports=$TCP_REDIR_PORTS
[ "$udp_redir_ports" = "default" ] && udp_redir_ports=$UDP_REDIR_PORTS
[ -n "$(get_cache_var "ACL_${sid}_tcp_node")" ] && tcp_node=$(get_cache_var "ACL_${sid}_tcp_node")
[ -n "$(get_cache_var "ACL_${sid}_tcp_redir_port")" ] && tcp_port=$(get_cache_var "ACL_${sid}_tcp_redir_port")
[ -n "$(get_cache_var "ACL_${sid}_udp_node")" ] && udp_node=$(get_cache_var "ACL_${sid}_udp_node")
[ -n "$(get_cache_var "ACL_${sid}_udp_redir_port")" ] && udp_port=$(get_cache_var "ACL_${sid}_udp_redir_port")
[ -n "$(get_cache_var "ACL_${sid}_dns_port")" ] && dns_redirect_port=$(get_cache_var "ACL_${sid}_dns_port")
[ -n "$tcp_node" ] && tcp_node_remark=$(config_n_get $tcp_node remarks)
[ -n "$udp_node" ] && udp_node_remark=$(config_n_get $udp_node remarks)
use_shunt_tcp=0
use_shunt_udp=0
[ -n "$tcp_node" ] && [ "$(config_n_get $tcp_node protocol)" = "_shunt" ] && use_shunt_tcp=1
[ -n "$udp_node" ] && [ "$(config_n_get $udp_node protocol)" = "_shunt" ] && use_shunt_udp=1
[ "${use_global_config}" = "1" ] && {
tcp_node_remark=$(config_n_get $TCP_NODE remarks)
udp_node_remark=$(config_n_get $UDP_NODE remarks)
use_direct_list=${USE_DIRECT_LIST}
use_proxy_list=${USE_PROXY_LIST}
use_block_list=${USE_BLOCK_LIST}
use_gfw_list=${USE_GFW_LIST}
chn_list=${CHN_LIST}
tcp_proxy_mode=${TCP_PROXY_MODE}
udp_proxy_mode=${UDP_PROXY_MODE}
use_shunt_tcp=${USE_SHUNT_TCP}
use_shunt_udp=${USE_SHUNT_UDP}
dns_redirect_port=${DNS_REDIRECT_PORT}
black_set_name=${NFTSET_BLACK}
black6_set_name=${NFTSET_BLACK6}
gfw_set_name=${NFTSET_GFW}
gfw6_set_name=${NFTSET_GFW6}
shunt_set_name=${NFTSET_SHUNT}
shunt6_set_name=${NFTSET_SHUNT6}
}
_acl_list=${TMP_ACL_PATH}/${sid}/source_list
for i in $(cat $_acl_list); do
local _ipt_source _ipv4
local msg
if [ -n "${interface}" ]; then
. /lib/functions/network.sh
local gateway device
network_get_gateway gateway "${interface}"
network_get_device device "${interface}"
[ -z "${device}" ] && device="${interface}"
_ipt_source="iifname ${device} "
msg="源接口【${device}】,"
else
msg="源接口【所有】,"
fi
if [ -n "$(echo ${i} | grep '^iprange:')" ]; then
_iprange=$(echo ${i} | sed 's#iprange:##g')
_ipt_source=$(factor ${_iprange} "${_ipt_source}ip saddr")
msg="${msg}IP range【${_iprange}】,"
_ipv4="1"
unset _iprange
elif [ -n "$(echo ${i} | grep '^ipset:')" ]; then
_ipset=$(echo ${i} | sed 's#ipset:##g')
_ipt_source="${_ipt_source}ip saddr @${_ipset}"
msg="${msg}NFTset【${_ipset}】,"
unset _ipset
elif [ -n "$(echo ${i} | grep '^ip:')" ]; then
_ip=$(echo ${i} | sed 's#ip:##g')
_ipt_source=$(factor ${_ip} "${_ipt_source}ip saddr")
msg="${msg}IP【${_ip}】,"
_ipv4="1"
unset _ip
elif [ -n "$(echo ${i} | grep '^mac:')" ]; then
_mac=$(echo ${i} | sed 's#mac:##g')
_ipt_source=$(factor ${_mac} "${_ipt_source}ether saddr")
msg="${msg}MAC【${_mac}】,"
unset _mac
elif [ -n "$(echo ${i} | grep '^any')" ]; then
msg="${msg}所有设备,"
else
continue
fi
msg="【$remarks】,${msg}"
[ "$tcp_no_redir_ports" != "disable" ] && {
if ! has_1_65535 "$tcp_no_redir_ports"; then
nft "add rule $NFTABLE_NAME $nft_prerouting_chain ${_ipt_source} ip protocol tcp $(factor $tcp_no_redir_ports "tcp dport") counter return comment \"$remarks\""
[ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ${_ipt_source} meta l4proto tcp $(factor $tcp_no_redir_ports "tcp dport") counter return comment \"$remarks\""
echolog " - ${msg}不代理 TCP 端口[${tcp_no_redir_ports}]"
else
#结束时会return,无需加多余的规则。
unset tcp_port
echolog " - ${msg}不代理所有 TCP 端口"
fi
}
[ "$udp_no_redir_ports" != "disable" ] && {
if ! has_1_65535 "$udp_no_redir_ports"; then
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_no_redir_ports "udp dport") counter return comment \"$remarks\""
[ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_no_redir_ports "udp dport") counter return comment \"$remarks\"" 2>/dev/null
echolog " - ${msg}不代理 UDP 端口[${udp_no_redir_ports}]"
else
#结束时会return,无需加多余的规则。
unset udp_port
echolog " - ${msg}不代理所有 UDP 端口"
fi
}
local dns_redirect
[ $(config_t_get global dns_redirect "1") = "1" ] && dns_redirect=53
if ([ -n "$tcp_port" ] && [ -n "${tcp_proxy_mode}" ]) || ([ -n "$udp_port" ] && [ -n "${udp_proxy_mode}" ]); then
[ "${use_proxy_list}" = "1" ] && {
[ "${use_global_config}" = "0" ] && {
black_set_name="passwall_${sid}_black"
black6_set_name="passwall_${sid}_black6"
gen_nftset $black_set_name ipv4_addr "2d" 0
gen_nftset $black6_set_name ipv6_addr "2d" 0
}
}
[ "${use_gfw_list}" = "1" ] && {
[ "${use_global_config}" = "0" ] && {
gfw_set_name="passwall_${sid}_gfw"
gfw6_set_name="passwall_${sid}_gfw6"
gen_nftset $gfw_set_name ipv4_addr "2d" 0
gen_nftset $gfw6_set_name ipv6_addr "2d" 0
}
}
[ "${use_shunt_tcp}" = "1" -o "${use_shunt_udp}" = "1" ] && {
[ "${use_global_config}" = "0" ] && {
shunt_set_name="passwall_${sid}_shunt"
shunt6_set_name="passwall_${sid}_shunt6"
gen_nftset $shunt_set_name ipv4_addr "2d" 0
gen_nftset $shunt6_set_name ipv6_addr "2d" 0
}
}
[ -n "${dns_redirect_port}" ] && dns_redirect=${dns_redirect_port}
else
[ -n "${DIRECT_DNSMASQ_PORT}" ] && dns_redirect=${DIRECT_DNSMASQ_PORT}
fi
if [ -n "${dns_redirect}" ]; then
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} udp dport 53 counter return comment \"$remarks\""
[ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} udp dport 53 counter return comment \"$remarks\""
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp ${_ipt_source} tcp dport 53 counter return comment \"$remarks\""
[ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} tcp dport 53 counter return comment \"$remarks\""
#nft "add rule $NFTABLE_NAME PSW_DNS ip protocol udp ${_ipt_source} udp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\""
#nft "add rule $NFTABLE_NAME PSW_DNS ip protocol tcp ${_ipt_source} tcp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\""
nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto udp ${_ipt_source} udp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\""
nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto tcp ${_ipt_source} tcp dport 53 counter redirect to :${dns_redirect} comment \"$remarks\""
[ -z "$(get_cache_var "ACL_${sid}_tcp_default")" ] && echolog " - ${msg}使用与全局配置不相同节点,已将DNS强制重定向到专用 DNS 服务器。"
fi
[ -n "$tcp_port" -o -n "$udp_port" ] && {
[ "${use_block_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ${_ipt_source} ip daddr @$NFTSET_BLOCK counter drop comment \"$remarks\""
[ "${use_block_list}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ${_ipt_source} ip daddr @$NFTSET_BLOCK counter drop comment \"$remarks\""
[ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ${_ipt_source} ip daddr @$NFTSET_WHITE counter return comment \"$remarks\""
[ "${use_direct_list}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ${_ipt_source} ip daddr @$NFTSET_WHITE counter return comment \"$remarks\""
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
[ "${use_block_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ${_ipt_source} ip6 daddr @$NFTSET_BLOCK6 counter drop comment \"$remarks\""
[ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ${_ipt_source} ip6 daddr @$NFTSET_WHITE6 counter return comment \"$remarks\""
}
[ "$tcp_proxy_drop_ports" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"$remarks\"" 2>/dev/null
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$black6_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$gfw6_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\"" 2>/dev/null
[ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip6 daddr @$shunt6_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") counter drop comment \"$remarks\"" 2>/dev/null
}
nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr $FAKE_IP counter drop comment \"$remarks\""
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$black_set_name counter drop comment \"$remarks\""
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$gfw_set_name counter drop comment \"$remarks\""
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\""
[ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") ip daddr @$shunt_set_name counter drop comment \"$remarks\""
[ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} $(factor $tcp_proxy_drop_ports "tcp dport") counter drop comment \"$remarks\""
echolog " - ${msg}屏蔽代理 TCP 端口[${tcp_proxy_drop_ports}]"
}
[ "$udp_proxy_drop_ports" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"$remarks\"" 2>/dev/null
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$black6_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$gfw6_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\"" 2>/dev/null
[ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip6 daddr @$shunt6_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") counter drop comment \"$remarks\"" 2>/dev/null
}
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr $FAKE_IP counter drop comment \"$remarks\"" 2>/dev/null
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$black_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$gfw_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "counter drop") comment \"$remarks\"" 2>/dev/null
[ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") ip daddr @$shunt_set_name counter drop comment \"$remarks\"" 2>/dev/null
[ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_proxy_drop_ports "udp dport") counter drop comment \"$remarks\"" 2>/dev/null
echolog " - ${msg}屏蔽代理 UDP 端口[${udp_proxy_drop_ports}]"
}
}
[ -n "$tcp_port" ] && {
if [ -n "${tcp_proxy_mode}" ]; then
msg2="${msg}使用 TCP 节点[$tcp_node_remark]"
if [ -n "${is_tproxy}" ]; then
msg2="${msg2}(TPROXY:${tcp_port})"
nft_chain="PSW_MANGLE"
nft_j="counter jump PSW_RULE"
else
msg2="${msg2}(REDIRECT:${tcp_port})"
nft_chain="PSW_NAT"
nft_j="$(REDIRECT $tcp_port)"
fi
[ "$accept_icmp" = "1" ] && {
[ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$NFTSET_WHITE counter return comment \"$remarks\""
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr $FAKE_IP $(REDIRECT) comment \"$remarks\""
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$black_set_name $(REDIRECT) comment \"$remarks\""
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$gfw_set_name $(REDIRECT) comment \"$remarks\""
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list}) comment \"$remarks\""
[ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} ip daddr @$shunt_set_name $(REDIRECT) comment \"$remarks\""
[ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} $(REDIRECT) comment \"$remarks\""
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ${_ipt_source} return comment \"$remarks\""
}
[ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
[ "${use_direct_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$NFTSET_WHITE6 counter return comment \"$remarks\"" 2>/dev/null
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr $FAKE_IP_6 $(REDIRECT) comment \"$remarks\"" 2>/dev/null
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$black6_set_name $(REDIRECT) comment \"$remarks\"" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$gfw6_set_name $(REDIRECT) comment \"$remarks\"" 2>/dev/null
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list}) comment \"$remarks\"" 2>/dev/null
[ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} ip6 daddr @$shunt6_set_name $(REDIRECT) comment \"$remarks\"" 2>/dev/null
[ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} $(REDIRECT) comment \"$remarks\"" 2>/dev/null
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ${_ipt_source} return comment \"$remarks\"" 2>/dev/null
}
nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} ip daddr $FAKE_IP ${nft_j} comment \"$remarks\""
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$black_set_name ${nft_j} comment \"$remarks\" "
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$gfw_set_name ${nft_j} comment \"$remarks\" "
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "${nft_j}") comment \"$remarks\" "
[ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip daddr @$shunt_set_name ${nft_j} comment \"$remarks\""
[ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ${nft_j} comment \"$remarks\""
[ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY4) comment \"$remarks\""
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"$remarks\""
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$black6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$gfw6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter jump PSW_RULE") comment \"$remarks\" "
[ "${use_shunt_tcp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") ip6 daddr @$shunt6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
[ "${tcp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(factor $tcp_redir_ports "tcp dport") counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY) comment \"$remarks\"" 2>/dev/null
}
else
msg2="${msg}不代理 TCP"
fi
echolog " - ${msg2}"
}
nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp ${_ipt_source} counter return comment \"$remarks\""
[ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ${_ipt_source} counter return comment \"$remarks\"" 2>/dev/null
[ -n "$udp_port" ] && {
if [ -n "${udp_proxy_mode}" ]; then
msg2="${msg}使用 UDP 节点[$udp_node_remark]"
msg2="${msg2}(TPROXY:${udp_port})"
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} ip daddr $FAKE_IP counter jump PSW_RULE comment \"$remarks\""
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$black_set_name counter jump PSW_RULE comment \"$remarks\""
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$gfw_set_name counter jump PSW_RULE comment \"$remarks\""
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${chn_list} "counter jump PSW_RULE") comment \"$remarks\""
[ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip daddr @$shunt_set_name counter jump PSW_RULE comment \"$remarks\""
[ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") counter jump PSW_RULE comment \"$remarks\""
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} $(REDIRECT $udp_port TPROXY4) comment \"$remarks\""
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"$remarks\""
[ "${use_proxy_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$black6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$gfw6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
[ "${chn_list}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${chn_list} "counter jump PSW_RULE") comment \"$remarks\"" 2>/dev/null
[ "${use_shunt_udp}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") ip6 daddr @$shunt6_set_name counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
[ "${udp_proxy_mode}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(factor $udp_redir_ports "udp dport") counter jump PSW_RULE comment \"$remarks\"" 2>/dev/null
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} $(REDIRECT $udp_port TPROXY) comment \"$remarks\"" 2>/dev/null
}
else
msg2="${msg}不代理 UDP"
fi
echolog " - ${msg2}"
}
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ${_ipt_source} counter return comment \"$remarks\""
[ "$_ipv4" != "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ${_ipt_source} counter return comment \"$remarks\"" 2>/dev/null
unset nft_chain nft_j _ipt_source msg msg2 _ipv4
done
unset enabled sid remarks sources use_global_config use_direct_list use_proxy_list use_block_list use_gfw_list chn_list tcp_proxy_mode udp_proxy_mode dns_redirect_port tcp_no_redir_ports udp_no_redir_ports tcp_proxy_drop_ports udp_proxy_drop_ports tcp_redir_ports udp_redir_ports tcp_node udp_node interface
unset tcp_port udp_port tcp_node_remark udp_node_remark _acl_list use_shunt_tcp use_shunt_udp dns_redirect
done
}
[ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ] && {
msg="【默认】,"
[ "$TCP_NO_REDIR_PORTS" != "disable" ] && {
nft "add rule $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return comment \"默认\""
if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]"
else
unset TCP_PROXY_MODE
echolog " - ${msg}不代理所有 TCP 端口"
fi
}
[ "$UDP_NO_REDIR_PORTS" != "disable" ] && {
nft "add $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return comment \"默认\""
nft "add $NFTABLE_NAME PSW_MANGLE_V6 counter meta l4proto udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return comment \"默认\""
if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]"
else
unset UDP_PROXY_MODE
echolog " - ${msg}不代理所有 UDP 端口"
fi
}
local DNS_REDIRECT
[ $(config_t_get global dns_redirect "1") = "1" ] && DNS_REDIRECT=53
if ([ -n "$TCP_NODE" ] && [ -n "${TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${UDP_PROXY_MODE}" ]); then
[ -n "${DNS_REDIRECT_PORT}" ] && DNS_REDIRECT=${DNS_REDIRECT_PORT}
else
[ -n "${DIRECT_DNSMASQ_PORT}" ] && DNS_REDIRECT=${DIRECT_DNSMASQ_PORT}
fi
if [ -n "${DNS_REDIRECT}" ]; then
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp udp dport 53 counter return comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp udp dport 53 counter return comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp tcp dport 53 counter return comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp tcp dport 53 counter return comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_DNS ip protocol udp udp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_DNS ip protocol tcp tcp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto udp udp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_DNS meta l4proto tcp tcp dport 53 counter redirect to :${DNS_REDIRECT} comment \"默认\""
fi
[ -n "${TCP_PROXY_MODE}" -o -n "${UDP_PROXY_MODE}" ] && {
[ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_BLOCK counter drop comment \"默认\""
[ "${USE_BLOCK_LIST}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_BLOCK counter drop comment \"默认\""
[ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_WHITE counter return comment \"默认\""
[ "${USE_DIRECT_LIST}" = "1" ] && [ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_WHITE counter return comment \"默认\""
[ "$PROXY_IPV6" == "1" ] && {
[ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_BLOCK6 counter drop comment \"默认\""
[ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_WHITE6 counter return comment \"默认\""
}
[ "$TCP_PROXY_DROP_PORTS" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_BLACK6 counter drop comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_GFW6 counter drop comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\""
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip6 daddr @$NFTSET_SHUNT6 counter drop comment \"默认\""
[ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop comment \"默认\""
}
nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr $FAKE_IP counter drop comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_BLACK counter drop comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_GFW counter drop comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\""
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") ip daddr @$NFTSET_SHUNT counter drop comment \"默认\""
[ "${TCP_PROXY_MODE}" != "disable" ] && nft "add $NFTABLE_NAME $nft_prerouting_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop comment \"默认\""
echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]"
}
[ "$UDP_PROXY_DROP_PORTS" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr $FAKE_IP_6 counter drop comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_BLACK6 counter drop comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_GFW6 counter drop comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\""
[ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip6 daddr @$NFTSET_SHUNT6 counter drop comment \"默认\""
[ "${UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop comment \"默认\""
}
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr $FAKE_IP counter drop comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_BLACK counter drop comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_GFW counter drop comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "counter drop") comment \"默认\""
[ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") ip daddr @$NFTSET_SHUNT counter drop comment \"默认\""
[ "${UDP_PROXY_MODE}" != "disable" ] && nft "add $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop comment \"默认\""
echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]"
}
}
# 加载TCP默认代理模式
if [ -n "${TCP_PROXY_MODE}" ]; then
[ -n "$TCP_NODE" ] && {
msg2="${msg}使用 TCP 节点[$(config_n_get $TCP_NODE remarks)]"
if [ -n "${is_tproxy}" ]; then
msg2="${msg2}(TPROXY:${TCP_REDIR_PORT})"
nft_chain="PSW_MANGLE"
nft_j="counter jump PSW_RULE"
else
msg2="${msg2}(REDIRECT:${TCP_REDIR_PORT})"
nft_chain="PSW_NAT"
nft_j="$(REDIRECT $TCP_REDIR_PORT)"
fi
[ "$accept_icmp" = "1" ] && {
[ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip daddr @$NFTSET_WHITE counter return comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr $FAKE_IP $(REDIRECT) comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_BLACK $(REDIRECT) comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_GFW $(REDIRECT) comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST}) comment \"默认\""
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp ip daddr @$NFTSET_SHUNT $(REDIRECT) comment \"默认\""
[ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp $(REDIRECT) comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip protocol icmp return comment \"默认\""
}
[ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && {
[ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip6 daddr @$NFTSET_WHITE6 counter return comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr $FAKE_IP_6 $(REDIRECT) comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_BLACK6 $(REDIRECT) comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_GFW6 $(REDIRECT) comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST}) comment \"默认\""
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 ip6 daddr @$NFTSET_SHUNT6 $(REDIRECT) comment \"默认\""
[ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 $(REDIRECT) comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT meta l4proto icmpv6 return comment \"默认\""
}
nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr $FAKE_IP ${nft_j} comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_BLACK ${nft_j} comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_GFW ${nft_j} comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "${nft_j}") comment \"默认\""
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip daddr @$NFTSET_SHUNT ${nft_j} comment \"默认\""
[ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j} comment \"默认\""
[ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(REDIRECT $TCP_REDIR_PORT TPROXY4) comment \"默认\""
nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp counter return comment \"默认\""
[ "$PROXY_IPV6" == "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_BLACK6 counter jump PSW_RULE comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_GFW6 counter jump PSW_RULE comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE") comment \"默认\""
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") ip6 daddr @$NFTSET_SHUNT6 counter jump PSW_RULE comment \"默认\""
[ "${TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp $(REDIRECT $TCP_REDIR_PORT TPROXY) comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp counter return comment \"默认\""
}
echolog " - ${msg2}"
}
fi
# 加载UDP默认代理模式
if [ -n "${UDP_PROXY_MODE}" ]; then
[ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ] && {
msg2="${msg}使用 UDP 节点[$(config_n_get $UDP_NODE remarks)](TPROXY:${UDP_REDIR_PORT})"
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp ip daddr $FAKE_IP counter jump PSW_RULE comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_BLACK counter jump PSW_RULE comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_GFW counter jump PSW_RULE comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE") comment \"默认\""
[ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") ip daddr @$NFTSET_SHUNT counter jump PSW_RULE comment \"默认\""
[ "${UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp $(REDIRECT $UDP_REDIR_PORT TPROXY4) comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp counter return comment \"默认\""
[ "$PROXY_IPV6" == "1" ] && {
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE comment \"默认\""
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_BLACK6 counter jump PSW_RULE comment \"默认\""
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_GFW6 counter jump PSW_RULE comment \"默认\""
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE") comment \"默认\""
[ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") ip6 daddr @$NFTSET_SHUNT6 counter jump PSW_RULE comment \"默认\""
[ "${UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp $(REDIRECT $UDP_REDIR_PORT TPROXY) comment \"默认\""
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp counter return comment \"默认\""
}
echolog " - ${msg2}"
}
fi
}
}
filter_haproxy() {
for item in ${haproxy_items}; do
local ip=$(get_host_ip ipv4 $(echo $item | awk -F ":" '{print $1}') 1)
insert_nftset $NFTSET_VPS "-1" $ip
done
echolog " - [$?]加入负载均衡的节点到nftset[$NFTSET_VPS]直连完成"
}
filter_vps_addr() {
for server_host in $@; do
local vps_ip4=$(get_host_ip "ipv4" ${server_host})
local vps_ip6=$(get_host_ip "ipv6" ${server_host})
[ -n "$vps_ip4" ] && insert_nftset $NFTSET_VPS "-1" $vps_ip4
[ -n "$vps_ip6" ] && insert_nftset $NFTSET_VPS6 "-1" $vps_ip6
done
}
filter_vpsip() {
insert_nftset $NFTSET_VPS "-1" $(uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([0-9]{1,3}[\.]){3}[0-9]{1,3}" | grep -v "^127\.0\.0\.1$" | sed -e "/^$/d")
echolog " - [$?]加入所有IPv4节点到nftset[$NFTSET_VPS]直连完成"
insert_nftset $NFTSET_VPS6 "-1" $(uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "/^$/d")
echolog " - [$?]加入所有IPv6节点到nftset[$NFTSET_VPS6]直连完成"
}
filter_server_port() {
local address=${1}
local port=${2}
local stream=${3}
stream=$(echo ${3} | tr 'A-Z' 'a-z')
local _is_tproxy
_is_tproxy=${is_tproxy}
[ "$stream" == "udp" ] && _is_tproxy="TPROXY"
for _ipt in 4 6; do
[ "$_ipt" == "4" ] && _ip_type=ip
[ "$_ipt" == "6" ] && _ip_type=ip6
nft "list chain $NFTABLE_NAME $nft_output_chain" 2>/dev/null | grep -q "${address}:${port}"
if [ $? -ne 0 ]; then
nft "insert rule $NFTABLE_NAME $nft_output_chain meta l4proto $stream $_ip_type daddr $address $stream dport $port return comment \"${address}:${port}\"" 2>/dev/null
fi
done
}
filter_node() {
local node=${1}
local stream=${2}
if [ -n "$node" ]; then
local address=$(config_n_get $node address)
local port=$(config_n_get $node port)
[ -z "$address" ] && [ -z "$port" ] && {
return 1
}
filter_server_port $address $port $stream
filter_server_port $address $port $stream
fi
}
filter_direct_node_list() {
[ ! -s "$TMP_PATH/direct_node_list" ] && return
for _node_id in $(cat $TMP_PATH/direct_node_list | awk '!seen[$0]++'); do
filter_node "$_node_id" TCP
filter_node "$_node_id" UDP
unset _node_id
done
}
add_firewall_rule() {
echolog "开始加载防火墙规则..."
gen_nft_tables
gen_nftset $NFTSET_VPS ipv4_addr 0 0
gen_nftset $NFTSET_GFW ipv4_addr "2d" 0
gen_nftset $NFTSET_LOCAL ipv4_addr 0 "-1"
gen_nftset $NFTSET_LAN ipv4_addr 0 "-1" $(gen_lanlist)
if [ -f $RULES_PATH/chnroute.nft ] && [ -s $RULES_PATH/chnroute.nft ] && [ $(awk 'END{print NR}' $RULES_PATH/chnroute.nft) -ge 8 ]; then
#echolog "使用缓存加载chnroute..."
nft -f $RULES_PATH/chnroute.nft
else
gen_nftset $NFTSET_CHN ipv4_addr "2d" 0 $(cat $RULES_PATH/chnroute | tr -s '\n' | grep -v "^#")
fi
gen_nftset $NFTSET_BLACK ipv4_addr "2d" 0
gen_nftset $NFTSET_WHITE ipv4_addr "2d" 0
gen_nftset $NFTSET_BLOCK ipv4_addr "2d" 0
gen_nftset $NFTSET_SHUNT ipv4_addr "2d" 0
gen_nftset $NFTSET_VPS6 ipv6_addr 0 0
gen_nftset $NFTSET_GFW6 ipv6_addr "2d" 0
gen_nftset $NFTSET_LOCAL6 ipv6_addr 0 "-1"
gen_nftset $NFTSET_LAN6 ipv6_addr 0 "-1" $(gen_lanlist_6)
if [ -f $RULES_PATH/chnroute6.nft ] && [ -s $RULES_PATH/chnroute6.nft ] && [ $(awk 'END{print NR}' $RULES_PATH/chnroute6.nft) -ge 8 ]; then
#echolog "使用缓存加载chnroute6..."
nft -f $RULES_PATH/chnroute6.nft
else
gen_nftset $NFTSET_CHN6 ipv6_addr "2d" 0 $(cat $RULES_PATH/chnroute6 | tr -s '\n' | grep -v "^#")
fi
gen_nftset $NFTSET_BLACK6 ipv6_addr "2d" 0
gen_nftset $NFTSET_WHITE6 ipv6_addr "2d" 0
gen_nftset $NFTSET_BLOCK6 ipv6_addr "2d" 0
gen_nftset $NFTSET_SHUNT6 ipv6_addr "2d" 0
#导入规则列表、分流规则中的IP列表
local USE_SHUNT_NODE=0
local USE_PROXY_LIST_ALL=${USE_PROXY_LIST}
local USE_DIRECT_LIST_ALL=${USE_DIRECT_LIST}
local USE_BLOCK_LIST_ALL=${USE_BLOCK_LIST}
local _TCP_NODE=$(config_t_get global tcp_node)
local _UDP_NODE=$(config_t_get global udp_node)
local USE_GEOVIEW=$(config_t_get global_rules enable_geoview)
[ -n "$_TCP_NODE" ] && [ "$(config_n_get $_TCP_NODE protocol)" = "_shunt" ] && USE_SHUNT_TCP=1 && USE_SHUNT_NODE=1
[ -n "$_UDP_NODE" ] && [ "$(config_n_get $_UDP_NODE protocol)" = "_shunt" ] && USE_SHUNT_UDP=1 && USE_SHUNT_NODE=1
[ "$_UDP_NODE" = "tcp" ] && USE_SHUNT_UDP=$USE_SHUNT_TCP
for acl_section in $(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do
[ "$(config_n_get $acl_section enabled)" != "1" ] && continue
[ "$(config_n_get $acl_section use_global_config 0)" != "1" ] && {
[ "$(config_n_get $acl_section use_direct_list 1)" = "1" ] && USE_PROXY_LIST_ALL=1
[ "$(config_n_get $acl_section use_proxy_list 1)" = "1" ] && USE_DIRECT_LIST_ALL=1
[ "$(config_n_get $acl_section use_block_list 1)" = "1" ] && USE_BLOCK_LIST_ALL=1
}
for _node in $(config_n_get $acl_section tcp_node) $(config_n_get $acl_section udp_node); do
local node_protocol=$(config_n_get $_node protocol)
[ "$node_protocol" = "_shunt" ] && { USE_SHUNT_NODE=1; break; }
done
done
#直连列表
[ "$USE_DIRECT_LIST_ALL" = "1" ] && {
insert_nftset $NFTSET_WHITE "0" $(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_WHITE6 "0" $(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
[ "$USE_GEOVIEW" = "1" ] && {
local GEOIP_CODE=$(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
insert_nftset $NFTSET_WHITE "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_WHITE6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
echolog " - [$?]解析并加入[直连列表] GeoIP 到 NFTSET 完成"
fi
}
}
#代理列表
[ "$USE_PROXY_LIST_ALL" = "1" ] && {
insert_nftset $NFTSET_BLACK "0" $(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_BLACK6 "0" $(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
[ "$USE_GEOVIEW" = "1" ] && {
local GEOIP_CODE=$(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
insert_nftset $NFTSET_BLACK "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_BLACK6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
echolog " - [$?]解析并加入[代理列表] GeoIP 到 NFTSET 完成"
fi
}
}
#屏蔽列表
[ "$USE_BLOCK_LIST_ALL" = "1" ] && {
insert_nftset $NFTSET_BLOCK "0" $(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_BLOCK6 "0" $(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
[ "$USE_GEOVIEW" = "1" ] && {
local GEOIP_CODE=$(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
insert_nftset $NFTSET_BLOCK "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_BLOCK6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
echolog " - [$?]解析并加入[屏蔽列表] GeoIP 到 NFTSET 完成"
fi
}
}
#分流列表
[ "$USE_SHUNT_NODE" = "1" ] && {
local GEOIP_CODE=""
local shunt_ids=$(uci show $CONFIG | grep "=shunt_rules" | awk -F '.' '{print $2}' | awk -F '=' '{print $1}')
for shunt_id in $shunt_ids; do
insert_nftset $NFTSET_SHUNT "0" $(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_SHUNT6 "0" $(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -v "^#" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
[ "$USE_GEOVIEW" = "1" ] && {
local geoip_code=$(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
[ -n "$geoip_code" ] && GEOIP_CODE="${GEOIP_CODE:+$GEOIP_CODE,}$geoip_code"
}
done
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
insert_nftset $NFTSET_SHUNT "0" $(get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}")
insert_nftset $NFTSET_SHUNT6 "0" $(get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
echolog " - [$?]解析并加入[分流节点] GeoIP 到 NFTSET 完成"
fi
}
insert_nftset $NFTSET_LOCAL "-1" $(ip address show | grep -w "inet" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/ /\n/g")
insert_nftset $NFTSET_LOCAL6 "-1" $(ip address show | grep -w "inet6" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/ /\n/g")
# 忽略特殊IP段
local lan_ifname lan_ip
lan_ifname=$(uci -q -p /tmp/state get network.lan.ifname)
[ -n "$lan_ifname" ] && {
lan_ip=$(ip address show $lan_ifname | grep -w "inet" | awk '{print $2}')
lan_ip6=$(ip address show $lan_ifname | grep -w "inet6" | awk '{print $2}')
#echolog "本机IPv4网段互访直连:${lan_ip}"
#echolog "本机IPv6网段互访直连:${lan_ip6}"
[ -n "$lan_ip" ] && insert_nftset $NFTSET_LAN "-1" $(echo $lan_ip | sed -e "s/ /\n/g")
[ -n "$lan_ip6" ] && insert_nftset $NFTSET_LAN6 "-1" $(echo $lan_ip6 | sed -e "s/ /\n/g")
}
[ -n "$ISP_DNS" ] && {
#echolog "处理 ISP DNS 例外..."
for ispip in $ISP_DNS; do
insert_nftset $NFTSET_WHITE 0 $ispip
echolog " - [$?]追加ISP IPv4 DNS到白名单:${ispip}"
done
}
[ -n "$ISP_DNS6" ] && {
#echolog "处理 ISP IPv6 DNS 例外..."
for ispip6 in $ISP_DNS6; do
insert_nftset $NFTSET_WHITE6 0 $ispip6
echolog " - [$?]追加ISP IPv6 DNS到白名单:${ispip6}"
done
}
# 过滤所有节点IP
filter_vpsip > /dev/null 2>&1 &
# filter_haproxy > /dev/null 2>&1 &
# Prevent some conditions
filter_vps_addr $(config_n_get $TCP_NODE address) $(config_n_get $UDP_NODE address) > /dev/null 2>&1 &
filter_vps_addr $(config_n_get $TCP_NODE download_address) $(config_n_get $UDP_NODE download_address) > /dev/null 2>&1 &
accept_icmp=$(config_t_get global_forwarding accept_icmp 0)
accept_icmpv6=$(config_t_get global_forwarding accept_icmpv6 0)
if [ "${TCP_PROXY_WAY}" = "redirect" ]; then
unset is_tproxy
nft_prerouting_chain="PSW_NAT"
nft_output_chain="PSW_OUTPUT_NAT"
elif [ "${TCP_PROXY_WAY}" = "tproxy" ]; then
is_tproxy="TPROXY"
nft_prerouting_chain="PSW_MANGLE"
nft_output_chain="PSW_OUTPUT_MANGLE"
fi
nft "add chain $NFTABLE_NAME PSW_DIVERT"
nft "flush chain $NFTABLE_NAME PSW_DIVERT"
nft "add rule $NFTABLE_NAME PSW_DIVERT meta l4proto tcp socket transparent 1 mark set 1 counter accept"
nft "add chain $NFTABLE_NAME PSW_DNS"
nft "flush chain $NFTABLE_NAME PSW_DNS"
if [ $(config_t_get global dns_redirect "1") = "0" ]; then
#Only hijack when dest address is local IP
nft "insert rule $NFTABLE_NAME dstnat ip daddr @${NFTSET_LOCAL} jump PSW_DNS"
nft "insert rule $NFTABLE_NAME dstnat ip6 daddr @${NFTSET_LOCAL6} jump PSW_DNS"
else
nft "insert rule $NFTABLE_NAME dstnat jump PSW_DNS"
fi
# for ipv4 ipv6 tproxy mark
nft "add chain $NFTABLE_NAME PSW_RULE"
nft "flush chain $NFTABLE_NAME PSW_RULE"
nft "add rule $NFTABLE_NAME PSW_RULE meta mark set ct mark counter"
nft "add rule $NFTABLE_NAME PSW_RULE meta mark 1 counter return"
nft "add rule $NFTABLE_NAME PSW_RULE tcp flags &(fin|syn|rst|ack) == syn meta mark set mark and 0x0 xor 0x1 counter"
nft "add rule $NFTABLE_NAME PSW_RULE meta l4proto udp ct state new meta mark set mark and 0x0 xor 0x1 counter"
nft "add rule $NFTABLE_NAME PSW_RULE ct mark set mark counter"
#ipv4 tproxy mode and udp
nft "add chain $NFTABLE_NAME PSW_MANGLE"
nft "flush chain $NFTABLE_NAME PSW_MANGLE"
nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_LAN counter return"
nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr @$NFTSET_VPS counter return"
nft "add chain $NFTABLE_NAME PSW_OUTPUT_MANGLE"
nft "flush chain $NFTABLE_NAME PSW_OUTPUT_MANGLE"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_LAN counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_VPS counter return"
[ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_BLOCK counter drop"
[ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip daddr @$NFTSET_WHITE counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE meta mark 0xff counter return"
# jump chains
nft "add rule $NFTABLE_NAME mangle_prerouting ip protocol udp counter jump PSW_MANGLE"
[ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME mangle_prerouting ip protocol tcp counter jump PSW_MANGLE"
insert_rule_before "$NFTABLE_NAME" "mangle_prerouting" "PSW_MANGLE" "counter jump PSW_DIVERT"
#ipv4 tcp redirect mode
[ -z "${is_tproxy}" ] && {
nft "add chain $NFTABLE_NAME PSW_NAT"
nft "flush chain $NFTABLE_NAME PSW_NAT"
nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_LAN counter return"
nft "add rule $NFTABLE_NAME PSW_NAT ip daddr @$NFTSET_VPS counter return"
nft "add rule $NFTABLE_NAME dstnat ip protocol tcp counter jump PSW_NAT"
nft "add chain $NFTABLE_NAME PSW_OUTPUT_NAT"
nft "flush chain $NFTABLE_NAME PSW_OUTPUT_NAT"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_LAN counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_VPS counter return"
[ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_BLOCK counter drop"
[ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT ip daddr @$NFTSET_WHITE counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_NAT meta mark 0xff counter return"
}
#icmp ipv6-icmp redirect
if [ "$accept_icmp" = "1" ]; then
nft "add chain $NFTABLE_NAME PSW_ICMP_REDIRECT"
nft "flush chain $NFTABLE_NAME PSW_ICMP_REDIRECT"
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip daddr @$NFTSET_LAN counter return"
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip daddr @$NFTSET_VPS counter return"
[ "$accept_icmpv6" = "1" ] && {
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip6 daddr @$NFTSET_LAN6 counter return"
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT ip6 daddr @$NFTSET_VPS6 counter return"
}
nft "add rule $NFTABLE_NAME dstnat meta l4proto {icmp,icmpv6} counter jump PSW_ICMP_REDIRECT"
nft "add rule $NFTABLE_NAME nat_output meta l4proto {icmp,icmpv6} counter jump PSW_ICMP_REDIRECT"
fi
WAN_IP=$(get_wan_ip)
if [ -n "${WAN_IP}" ]; then
[ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_NAT ip daddr ${WAN_IP} counter return comment \"WAN_IP_RETURN\""
nft "add rule $NFTABLE_NAME PSW_MANGLE ip daddr ${WAN_IP} counter return comment \"WAN_IP_RETURN\""
echolog " - [$?]追加WAN IP到nftables:${WAN_IP}"
fi
unset WAN_IP
ip rule add fwmark 1 lookup 100
ip route add local 0.0.0.0/0 dev lo table 100
#ipv6 tproxy mode and udp
nft "add chain $NFTABLE_NAME PSW_MANGLE_V6"
nft "flush chain $NFTABLE_NAME PSW_MANGLE_V6"
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_LAN6 counter return"
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr @$NFTSET_VPS6 counter return"
nft "add chain $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6"
nft "flush chain $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_LAN6 counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_VPS6 counter return"
[ "${USE_BLOCK_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_BLOCK6 counter drop"
[ "${USE_DIRECT_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 ip6 daddr @$NFTSET_WHITE6 counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta mark 0xff counter return"
[ -n "$IPT_APPEND_DNS" ] && {
local local_dns dns_address dns_port
for local_dns in $(echo $IPT_APPEND_DNS | tr ',' ' '); do
dns_address=$(echo "$local_dns" | sed -E 's/(@|\[)?([0-9a-fA-F:.]+)(@|#|$).*/\2/')
dns_port=$(echo "$local_dns" | sed -nE 's/.*#([0-9]+)$/\1/p')
if echo "$dns_address" | grep -q -v ':'; then
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr ${dns_address} $(factor ${dns_port:-53} "udp dport") counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol tcp ip daddr ${dns_address} $(factor ${dns_port:-53} "tcp dport") counter return"
echolog " - [$?]追加直连DNS到nftables:${dns_address}:${dns_port:-53}"
else
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr ${dns_address} $(factor ${dns_port:-53} "udp dport") counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr ${dns_address} $(factor ${dns_port:-53} "tcp dport") counter return"
echolog " - [$?]追加直连DNS到nftables:[${dns_address}]:${dns_port:-53}"
fi
done
}
# jump chains
[ "$PROXY_IPV6" == "1" ] && {
nft "add rule $NFTABLE_NAME mangle_prerouting meta nfproto {ipv6} counter jump PSW_MANGLE_V6"
nft "add rule $NFTABLE_NAME mangle_output meta nfproto {ipv6} counter jump PSW_OUTPUT_MANGLE_V6 comment \"PSW_OUTPUT_MANGLE\""
WAN6_IP=$(get_wan6_ip)
[ -n "${WAN6_IP}" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 ip6 daddr ${WAN6_IP} counter return comment \"WAN6_IP_RETURN\""
unset WAN6_IP
ip -6 rule add fwmark 1 table 100
ip -6 route add local ::/0 dev lo table 100
}
[ "$TCP_UDP" = "1" ] && [ -z "$UDP_NODE" ] && UDP_NODE=$TCP_NODE
[ "$ENABLED_DEFAULT_ACL" == 1 ] && {
msg="【路由器本机】,"
[ "$TCP_NO_REDIR_PORTS" != "disable" ] && {
nft "add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp $(factor $TCP_NO_REDIR_PORTS "tcp dport") counter return"
if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]"
else
unset LOCALHOST_TCP_PROXY_MODE
echolog " - ${msg}不代理所有 TCP 端口"
fi
}
[ "$UDP_NO_REDIR_PORTS" != "disable" ] && {
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return"
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp $(factor $UDP_NO_REDIR_PORTS "udp dport") counter return"
if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]"
else
unset LOCALHOST_UDP_PROXY_MODE
echolog " - ${msg}不代理所有 UDP 端口"
fi
}
if ([ -n "$TCP_NODE" ] && [ -n "${LOCALHOST_TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${LOCALHOST_UDP_PROXY_MODE}" ]); then
[ -n "$DNS_REDIRECT_PORT" ] && {
nft "add rule $NFTABLE_NAME nat_output ip protocol udp oif lo udp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\""
nft "add rule $NFTABLE_NAME nat_output ip protocol tcp oif lo tcp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\""
nft "add rule $NFTABLE_NAME nat_output meta l4proto udp oif lo udp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\""
nft "add rule $NFTABLE_NAME nat_output meta l4proto tcp oif lo tcp dport 53 counter redirect to :$DNS_REDIRECT_PORT comment \"PSW_DNS\""
}
fi
[ -n "${LOCALHOST_TCP_PROXY_MODE}" -o -n "${LOCALHOST_UDP_PROXY_MODE}" ] && {
[ "$TCP_PROXY_DROP_PORTS" != "disable" ] && {
nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr $FAKE_IP $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop
[ "${USE_PROXY_LIST}" = "1" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_BLACK $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop
[ "${USE_GFW_LIST}" = "1" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_GFW $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop
[ "${CHN_LIST}" != "0" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_CHN $(factor $TCP_PROXY_DROP_PORTS "tcp dport") $(get_jump_ipt ${CHN_LIST} "counter drop")
[ "${USE_SHUNT_TCP}" = "1" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp ip daddr @$NFTSET_SHUNT $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop
[ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft add rule $NFTABLE_NAME $nft_output_chain ip protocol tcp $(factor $TCP_PROXY_DROP_PORTS "tcp dport") counter drop
echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]"
}
[ "$UDP_PROXY_DROP_PORTS" != "disable" ] && {
nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr $FAKE_IP $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop
[ "${USE_PROXY_LIST}" = "1" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_BLACK $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop
[ "${USE_GFW_LIST}" = "1" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_GFW $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop
[ "${CHN_LIST}" != "0" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_CHN $(factor $UDP_PROXY_DROP_PORTS "udp dport") $(get_jump_ipt ${CHN_LIST} "counter drop")
[ "${USE_SHUNT_UDP}" = "1" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_SHUNT $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop
[ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && nft add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE counter ip protocol udp $(factor $UDP_PROXY_DROP_PORTS "udp dport") counter drop
echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]"
}
}
# 加载路由器自身代理 TCP
if [ -n "$TCP_NODE" ]; then
_proxy_tcp_access() {
[ -n "${2}" ] || return 0
if echo "${2}" | grep -q -v ':'; then
nft "get element $NFTABLE_NAME $NFTSET_LAN {${2}}" &>/dev/null
[ $? -eq 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问"
return 0
}
if [ -z "${is_tproxy}" ]; then
nft insert rule $NFTABLE_NAME PSW_OUTPUT_NAT ip protocol tcp ip daddr ${2} tcp dport ${3} $(REDIRECT $TCP_REDIR_PORT)
else
nft insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol tcp ip daddr ${2} tcp dport ${3} counter jump PSW_RULE
nft insert rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp iif lo tcp dport ${3} ip daddr ${2} $(REDIRECT $TCP_REDIR_PORT TPROXY4) comment \"本机\"
fi
echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 TCP 转发链"
else
nft "get element $NFTABLE_NAME $NFTSET_LAN6 {${2}}" &>/dev/null
[ $? -eq 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问"
return 0
}
nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr ${2} tcp dport ${3} counter jump PSW_RULE"
nft "insert rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo tcp dport ${3} ip6 daddr ${2} $(REDIRECT $TCP_REDIR_PORT TPROXY6) comment \"本机\""
echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 TCP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!"
fi
}
[ "$TCP_PROXY_DNS" == 1 ] && hosts_foreach REMOTE_DNS _proxy_tcp_access 53
[ "$accept_icmp" = "1" ] && {
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr $FAKE_IP counter redirect"
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_BLACK counter redirect"
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_GFW counter redirect"
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_CHN $(get_jump_ipt ${CHN_LIST})"
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp ip daddr @$NFTSET_SHUNT counter redirect"
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp counter redirect"
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo ip protocol icmp counter return"
}
[ "$accept_icmpv6" = "1" ] && {
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr $FAKE_IP_6 counter redirect"
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_BLACK6 counter redirect"
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_GFW6 counter redirect"
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_CHN6 $(get_jump_ipt ${CHN_LIST})"
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 ip6 daddr @$NFTSET_SHUNT6 counter redirect"
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 counter redirect"
nft "add rule $NFTABLE_NAME PSW_ICMP_REDIRECT oif lo meta l4proto icmpv6 counter return"
}
if [ -n "${is_tproxy}" ]; then
nft_chain="PSW_OUTPUT_MANGLE"
nft_j="counter jump PSW_RULE"
else
nft_chain="PSW_OUTPUT_NAT"
nft_j="$(REDIRECT $TCP_REDIR_PORT)"
fi
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && {
nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr $FAKE_IP ${nft_j}"
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_BLACK $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}"
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_GFW $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}"
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_CHN $(factor $TCP_REDIR_PORTS "tcp dport") $(get_jump_ipt ${CHN_LIST} "${nft_j}")"
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp ip daddr @$NFTSET_SHUNT $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}"
[ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME $nft_chain ip protocol tcp $(factor $TCP_REDIR_PORTS "tcp dport") ${nft_j}"
[ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp iif lo $(REDIRECT $TCP_REDIR_PORT TPROXY4) comment \"本机\""
}
[ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol tcp iif lo counter return comment \"本机\""
[ -n "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME mangle_output ip protocol tcp counter jump PSW_OUTPUT_MANGLE comment \"PSW_OUTPUT_MANGLE\""
[ -z "${is_tproxy}" ] && nft "add rule $NFTABLE_NAME nat_output ip protocol tcp counter jump PSW_OUTPUT_NAT"
[ "$PROXY_IPV6" == "1" ] && {
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && {
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE"
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_BLACK6 $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_GFW6 $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE"
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_CHN6 $(factor $TCP_REDIR_PORTS "tcp dport") $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE")"
[ "${USE_SHUNT_TCP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp ip6 daddr @$NFTSET_SHUNT6 $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE"
[ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto tcp $(factor $TCP_REDIR_PORTS "tcp dport") counter jump PSW_RULE"
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo $(REDIRECT $TCP_REDIR_PORT TPROXY) comment \"本机\""
}
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo counter return comment \"本机\""
}
fi
# 加载路由器自身代理 UDP
if [ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ]; then
_proxy_udp_access() {
[ -n "${2}" ] || return 0
if echo "${2}" | grep -q -v ':'; then
nft "get element $NFTABLE_NAME $NFTSET_LAN {${2}}" &>/dev/null
[ $? == 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问"
return 0
}
nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr ${2} udp dport ${3} counter jump PSW_RULE"
nft "insert rule $NFTABLE_NAME PSW_MANGLE ip protocol udp iif lo ip daddr ${2} $(REDIRECT $UDP_REDIR_PORT TPROXY4) comment \"本机\""
echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 UDP 转发链"
else
nft "get element $NFTABLE_NAME $NFTSET_LAN6 {${2}}" &>/dev/null
[ $? == 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问"
return 0
}
nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr ${2} udp dport ${3} counter jump PSW_RULE"
nft "insert rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto tcp iif lo ip6 daddr ${2} $(REDIRECT $UDP_REDIR_PORT TPROXY6) comment \"本机\""
echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 UDP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!"
fi
}
[ -n "${UDP_PROXY_DNS}" ] && hosts_foreach REMOTE_DNS _proxy_udp_access 53
[ -n "${LOCALHOST_UDP_PROXY_MODE}" ] && {
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr $FAKE_IP counter jump PSW_RULE"
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_BLACK $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_GFW $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_CHN $(factor $UDP_REDIR_PORTS "udp dport") $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE")"
[ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp ip daddr @$NFTSET_SHUNT $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
[ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE ip protocol udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp iif lo $(REDIRECT $UDP_REDIR_PORT TPROXY4) comment \"本机\""
}
nft "add rule $NFTABLE_NAME PSW_MANGLE ip protocol udp iif lo counter return comment \"本机\""
nft "add rule $NFTABLE_NAME mangle_output ip protocol udp counter jump PSW_OUTPUT_MANGLE comment \"PSW_OUTPUT_MANGLE\""
[ "$PROXY_IPV6" == "1" ] && {
[ -n "${LOCALHOST_UDP_PROXY_MODE}" ] && {
nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr $FAKE_IP_6 counter jump PSW_RULE"
[ "${USE_PROXY_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_BLACK6 $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_GFW6 $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
[ "${CHN_LIST}" != "0" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_CHN6 $(factor $UDP_REDIR_PORTS "udp dport") $(get_jump_ipt ${CHN_LIST} "counter jump PSW_RULE")"
[ "${USE_SHUNT_UDP}" = "1" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp ip6 daddr @$NFTSET_SHUNT6 $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
[ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && nft "add rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 meta l4proto udp $(factor $UDP_REDIR_PORTS "udp dport") counter jump PSW_RULE"
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp iif lo $(REDIRECT $UDP_REDIR_PORT TPROXY) comment \"本机\""
}
nft "add rule $NFTABLE_NAME PSW_MANGLE_V6 meta l4proto udp iif lo counter return comment \"本机\""
}
fi
nft "add rule $NFTABLE_NAME mangle_output oif lo counter return comment \"PSW_OUTPUT_MANGLE\""
nft "add rule $NFTABLE_NAME mangle_output meta mark 1 counter return comment \"PSW_OUTPUT_MANGLE\""
}
# 加载ACLS
load_acl
[ -d "${TMP_IFACE_PATH}" ] && {
for iface in $(ls ${TMP_IFACE_PATH}); do
nft "insert rule $NFTABLE_NAME $nft_output_chain oif $iface counter return"
nft "insert rule $NFTABLE_NAME PSW_OUTPUT_MANGLE_V6 oif $iface counter return"
done
}
filter_direct_node_list > /dev/null 2>&1 &
echolog "防火墙规则加载完成!"
}
del_firewall_rule() {
for nft in "dstnat" "srcnat" "nat_output" "mangle_prerouting" "mangle_output"; do
local handles=$(nft -a list chain $NFTABLE_NAME ${nft} 2>/dev/null | grep -E "PSW_" | awk -F '# handle ' '{print$2}')
for handle in $handles; do
nft delete rule $NFTABLE_NAME ${nft} handle ${handle} 2>/dev/null
done
done
for handle in $(nft -a list chains | grep -E "chain PSW_" | grep -v "PSW_RULE" | awk -F '# handle ' '{print$2}'); do
nft delete chain $NFTABLE_NAME handle ${handle} 2>/dev/null
done
# Need to be removed at the end, otherwise it will show "Resource busy"
nft delete chain $NFTABLE_NAME handle $(nft -a list chains | grep -E "PSW_RULE" | awk -F '# handle ' '{print$2}') 2>/dev/null
ip rule del fwmark 1 lookup 100 2>/dev/null
ip route del local 0.0.0.0/0 dev lo table 100 2>/dev/null
ip -6 rule del fwmark 1 table 100 2>/dev/null
ip -6 route del local ::/0 dev lo table 100 2>/dev/null
destroy_nftset $NFTSET_LOCAL
destroy_nftset $NFTSET_LAN
destroy_nftset $NFTSET_VPS
destroy_nftset $NFTSET_SHUNT
#destroy_nftset $NFTSET_GFW
#destroy_nftset $NFTSET_CHN
destroy_nftset $NFTSET_BLACK
destroy_nftset $NFTSET_BLOCK
destroy_nftset $NFTSET_WHITE
destroy_nftset $NFTSET_LOCAL6
destroy_nftset $NFTSET_LAN6
destroy_nftset $NFTSET_VPS6
destroy_nftset $NFTSET_SHUNT6
#destroy_nftset $NFTSET_GFW6
#destroy_nftset $NFTSET_CHN6
destroy_nftset $NFTSET_BLACK6
destroy_nftset $NFTSET_BLOCK6
destroy_nftset $NFTSET_WHITE6
$DIR/app.sh echolog "删除 nftables 规则完成。"
}
flush_nftset() {
$DIR/app.sh echolog "清空 NFTSet。"
for _name in $(nft -a list sets | grep -E "passwall" | awk -F 'set ' '{print $2}' | awk '{print $1}'); do
destroy_nftset ${_name}
done
}
flush_table() {
nft flush table $NFTABLE_NAME
nft delete table $NFTABLE_NAME
}
flush_include() {
echo '#!/bin/sh' >$FWI
}
gen_include() {
flush_include
local nft_chain_file=$TMP_PATH/PSW_RULE.nft
echo '#!/usr/sbin/nft -f' > $nft_chain_file
nft list table $NFTABLE_NAME >> $nft_chain_file
local __nft=" "
__nft=$(cat <<- EOF
[ -z "\$(nft list chain $NFTABLE_NAME mangle_prerouting | grep PSW_DIVERT)" ] && nft -f ${nft_chain_file}
[ -z "${is_tproxy}" ] && {
PR_INDEX=\$(sh ${MY_PATH} RULE_LAST_INDEX "$NFTABLE_NAME" PSW_NAT WAN_IP_RETURN -1)
if [ \$PR_INDEX -ge 0 ]; then
WAN_IP=\$(sh ${MY_PATH} get_wan_ip)
[ ! -z "\${WAN_IP}" ] && nft "replace rule $NFTABLE_NAME PSW_NAT handle \$PR_INDEX ip daddr "\${WAN_IP}" counter return comment \"WAN_IP_RETURN\""
fi
}
PR_INDEX=\$(sh ${MY_PATH} RULE_LAST_INDEX "$NFTABLE_NAME" PSW_MANGLE WAN_IP_RETURN -1)
if [ \$PR_INDEX -ge 0 ]; then
WAN_IP=\$(sh ${MY_PATH} get_wan_ip)
[ ! -z "\${WAN_IP}" ] && nft "replace rule $NFTABLE_NAME PSW_MANGLE handle \$PR_INDEX ip daddr "\${WAN_IP}" counter return comment \"WAN_IP_RETURN\""
fi
[ "$PROXY_IPV6" == "1" ] && {
PR_INDEX=\$(sh ${MY_PATH} RULE_LAST_INDEX "$NFTABLE_NAME" PSW_MANGLE_V6 WAN6_IP_RETURN -1)
if [ \$PR_INDEX -ge 0 ]; then
WAN6_IP=\$(sh ${MY_PATH} get_wan6_ip)
[ ! -z "\${WAN_IP}" ] && nft "replace rule $NFTABLE_NAME PSW_MANGLE_V6 handle \$PR_INDEX ip6 daddr "\${WAN6_IP}" counter return comment \"WAN6_IP_RETURN\""
fi
}
EOF
)
cat <<-EOF >> $FWI
${__nft}
return 0
EOF
return 0
}
start() {
[ "$ENABLED_DEFAULT_ACL" == 0 -a "$ENABLED_ACLS" == 0 ] && return
add_firewall_rule
gen_include
}
stop() {
del_firewall_rule
[ $(config_t_get global flush_set "0") = "1" ] && {
uci -q delete ${CONFIG}.@global[0].flush_set
uci -q commit ${CONFIG}
#flush_table
flush_nftset
rm -rf /tmp/etc/passwall_tmp/singbox*
rm -rf /tmp/etc/passwall_tmp/dnsmasq*
}
flush_include
}
arg1=$1
shift
case $arg1 in
RULE_LAST_INDEX)
RULE_LAST_INDEX "$@"
;;
insert_rule_before)
insert_rule_before "$@"
;;
insert_rule_after)
insert_rule_after "$@"
;;
get_wan_ip)
get_wan_ip
;;
get_wan6_ip)
get_wan6_ip
;;
filter_direct_node_list)
filter_direct_node_list
;;
stop)
stop
;;
start)
start
;;
*) ;;
esac
|
294coder/Efficient-MIF | 3,551 | Pansharpening_Hyper_SR_Matlab_Test_Package/analysis_ref_unbatched_images.m | function res = analysis_ref_unbatched_images(dir_path, ratio, sensor, flag_cut_bounds, printout)
% data should be [0, max_range]
% e.g. wv3: max_range: 2047
if nargin == 3
flag_cut_bounds = 0;
printout = 1;
elseif nargin == 4
printout = 1;
end
sensor = lower(sensor);
if sensor == "wv3" || sensor == "wv2" || sensor == "qb"
const = 2047;
gt_key = 'gt';
if sensor == "wv3"
C = 8;
gt_path = "/Data2/ZiHanCao/datasets/pansharpening/wv3/reduced_examples/test_wv3_multiExm1.h5";
elseif sensor == "wv2"
C = 8;
gt_path = '/Data2/ZiHanCao/datasets/pansharpening/wv2/reduced_examples/test_wv2_multiExm1.h5';
else
C = 4;
gt_path = "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5";
end
elseif sensor == "gf2" || sensor == "gf"
const = 1023;
gt_key = 'gt';
C = 4;
gt_path = "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5";
elseif sensor == "cave" || sensor == "harvard"
const = 1;
gt_key = 'GT';
C = 31;
if sensor == "cave"
if ratio == 4
gt_path = "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5";
elseif ratio == 8
gt_path = "/Data2/ZiHanCao/datasets/HISI/new_cave/x8/test_cave(with_up)x8_rgb.h5";
end
else % % harvard
gt_path = '';
end
else
error(strcat(sensor, ' is not supported!'))
end
gts = h5read(gt_path, strcat('/', gt_key));
gts = permute(gts, [4, 2, 1, 3]);
bs = size(gts, 1);
addpath('./Tools')
addpath('./Quality_Indices/')
Q_block_size = 32;
thvalues = 0;
L = 11;
full_res = 0; % default to 0
if full_res
dim_cut = 21;
else
dim_cut = 30;
end
res = {};
sam = zeros(1, bs);
ergas = zeros(1, bs);
scc = zeros(1, bs);
qn = zeros(1, bs);
qave = zeros(1, bs);
psnr = zeros(1, bs);
ssim = zeros(1, bs);
for i = (0:bs - 1)
p = strcat(dir_path, "/", "output_mulExm_", string(i), ".mat");
data = load(p);
sr_key = fieldnames(data);
sr_key = sr_key{1};
sr1 = data.(string(sr_key));
% check the size
if size(sr1, 1) == 1 && length(size(sr1)) == 4
sr1 = squeeze(sr1);
end
if size(sr1, 3) ~= C && length(size(sr1)) == 3
sr1 = permute(sr1, [2, 3, 1]);
end
i = i + 1;
gt1 = squeeze(gts(i, :, :, :));
[q_avg_gt, sam_gt, ergas_gt, scc_gt, q_gt] = indexes_evaluation(sr1, gt1, ratio, L, Q_block_size, flag_cut_bounds, dim_cut, thvalues);
[psnr_gt, ssim_gt] = quality_assess(sr1 / const, gt1 / const);
sam(i) = sam_gt;
ergas(i) = ergas_gt;
scc(i) = scc_gt;
qn(i) = q_gt;
qave(i) = q_avg_gt;
psnr(i) = psnr_gt;
ssim(i) = ssim_gt;
if printout
fprintf("sample %d - sam: %f, ergas: %f, scc: %f, qn: %f, q_ave: %f, psnr: %f, ssim: %f \n", i, sam_gt, ergas_gt, scc_gt, q_gt, q_avg_gt, psnr_gt, ssim_gt)
end
end
res.sam = [mean(sam), std(sam)];
res.ergas = [mean(ergas), std(ergas)];
res.scc = [mean(scc), std(scc)];
res.qn = [mean(qn), std(qn)];
res.qave = [mean(qave), std(qave)];
res.psnr = [mean(psnr), std(psnr)];
res.ssim = [mean(ssim), std(ssim)];
end
|
294coder/Efficient-MIF | 4,692 | task_datasets/GF2.py | import torch
import torch.utils.data as data
import torchvision.transforms as T
import cv2
import numpy as np
import h5py
from typing import List, Tuple, Optional
class Identity:
def __call__(self, *args):
return args
class GF2Datasets(data.Dataset):
def __init__(
self,
d,
aug_prob=0.0,
hp=False,
hp_ksize=(5, 5),
norm_range=True,
full_res=False,
const=1023.0,
):
"""
:param d: h5py.File or dict or path
:param aug_prob: augmentation probability
:param hp: high pass for ms and pan. x = x - cv2.boxFilter(x)
:param hp_ksize: cv2.boxFiler kernel size
:param norm_range: normalize data range
"""
super(GF2Datasets, self).__init__()
# FIXME: should pass @path rather than @file which is h5py.File object to avoid can not be pickled error
if isinstance(d, (str, h5py.File)):
if isinstance(d, str):
d = h5py.File(d)
print(
"warning: when @file is a h5py.File object, it can not be pickled.",
"try to set DataLoader number_worker to 0",
)
if not full_res:
self.gt, self.ms, self.lms, self.pan = self.get_divided(d)
print("datasets shape:")
print("{:^20}{:^20}{:^20}{:^20}".format("pan", "ms", "lms", "gt"))
print(
"{:^20}{:^20}{:^20}{:^20}".format(
str(self.pan.shape),
str(self.ms.shape),
str(self.lms.shape),
str(self.gt.shape),
)
)
else:
self.ms, self.lms, self.pan = self.get_divided(d, True)
print("datasets shape:")
print("{:^20}{:^20}{:^20}".format("pan", "ms", "lms"))
print(
"{:^20}{:^20}{:^20}".format(
str(self.pan.shape), str(self.ms.shape), str(self.lms.shape)
)
)
self.size = self.ms.shape[0]
# highpass filter
self.hp = hp
self.hp_ksize = hp_ksize
if hp and hp_ksize is not None:
self.group_high_pass(hp_ksize)
# to tensor
if norm_range:
def norm_func(x):
# return torch.tensor(x) / 2047.
return torch.tensor(x, dtype=torch.float32) / const
else:
def norm_func(x):
return torch.tensor(x, dtype=torch.float32)
self.pan = norm_func(self.pan)
self.ms = norm_func(self.ms)
self.lms = norm_func(self.lms)
if not full_res:
self.gt = norm_func(self.gt)
# geometrical transformation
self.aug_prob = aug_prob
self.geo_trans = (
T.Compose(
[T.RandomVerticalFlip(p=aug_prob), T.RandomHorizontalFlip(p=aug_prob)]
)
if aug_prob != 0.0
else Identity()
)
@staticmethod
def get_divided(d, full_resolution=False):
if not full_resolution:
return (
np.asarray(d["gt"]),
np.asarray(d["ms"]),
np.asarray(d["lms"]),
np.asarray(d["pan"]),
)
else:
return (np.asarray(d["ms"]), np.asarray(d["lms"]), np.asarray(d["pan"]))
@staticmethod
def _get_high_pass(data, k_size):
for i, img in enumerate(data):
hp = cv2.boxFilter(img.transpose(1, 2, 0), -1, k_size)
if hp.ndim == 2:
hp = hp[..., np.newaxis]
data[i] = img - hp.transpose(2, 0, 1)
return data
def group_high_pass(self, k_size):
self.ms = self._get_high_pass(self.ms, k_size)
self.pan = self._get_high_pass(self.pan, k_size)
def aug_trans(self, *data):
data_list = []
seed = torch.random.seed()
for d in data:
torch.manual_seed(seed)
d = self.geo_trans(d)
data_list.append(d)
return data_list
def __getitem__(self, item):
if hasattr(self, "gt"):
tuple_data = (self.pan[item], self.ms[item], self.lms[item], self.gt[item])
else:
tuple_data = (self.pan[item], self.ms[item], self.lms[item])
return self.aug_trans(*tuple_data) if self.aug_prob != 0.0 else tuple_data
def __len__(self):
return self.size
def __repr__(self):
return (
f"num: {self.size} \n "
f"augmentation: {self.geo_trans} \n"
f"get high pass ms and pan: {self.hp} \n "
f"filter kernel size: {self.hp_ksize}"
)
|
294coder/Efficient-MIF | 6,408 | task_datasets/HISR.py | import time
from typing import Union
import matplotlib.pyplot as plt
import torch
import torch.utils.data as data
import torchvision.transforms as T
import h5py
import random
from typing import List, Tuple, Optional, Callable
def default_dataset_fn(*x):
return x[0]
class HISRDatasets(data.Dataset):
# FIXME: when use this Dataset, you should set num_works to 0 or it will raise unpickable error
def __init__(
self,
file: Union[h5py.File, str, dict],
aug_prob=0.0,
rgb_to_bgr=False,
full_res=False,
*,
dataset_fn=None
):
super(HISRDatasets, self).__init__()
# warning: you should not save file (h5py.File) in this class,
# or it will raise CAN NOT BE PICKLED error in multiprocessing
# FIXME: should pass @path rather than @file which is h5py.File object to avoid can not be pickled error
if isinstance(file, (str, h5py.File)):
if isinstance(file, str):
file = h5py.File(file)
print(
"warning: when @file is a h5py.File object, it can not be pickled.",
"try to set DataLoader number_worker to 0",
)
# checking dataset_fn type
if dataset_fn is not None:
if isinstance(dataset_fn, (list, tuple)):
def _apply_fn(tensor):
for fn in dataset_fn:
tensor = fn(tensor)
return tensor
self.dataset_fn = _apply_fn
elif isinstance(dataset_fn, Callable):
self.dataset_fn = dataset_fn
else: raise TypeError("dataset_fn should be a list of callable or a callable object")
else:
self.dataset_fn = default_dataset_fn
self.full_res = full_res
data_s= self._split_parts(
file, rgb_to_bgr=rgb_to_bgr, full=full_res
)
if len(data_s) == 4:
self.gt, self.lr_hsi, self.rgb, self.hsi_up = data_s
else:
self.lr_hsi, self.rgb, self.hsi_up = data_s
self.size = self.rgb.shape[-2:]
print("dataset shape:")
# print dataset info
if not full_res:
print("{:^20}{:^20}{:^20}{:^20}".format("lr_hsi", "hsi_up", "rgb", "gt"))
print(
"{:^20}{:^20}{:^20}{:^20}".format(
str(tuple(self.lr_hsi.shape)),
str(tuple(self.hsi_up.shape)),
str(tuple(self.rgb.shape)),
str(tuple(self.gt.shape)),
)
)
else:
print("{:^20}{:^20}{:^20}".format("lr_hsi", "hsi_up", "rgb"))
print(
"{:^20}{:^20}{:^20}".format(
str(tuple(self.lr_hsi.shape)),
str(tuple(self.hsi_up.shape)),
str(tuple(self.rgb.shape)),
)
)
# geometrical transformation
self.aug_prob = aug_prob
self.geo_trans = (
T.Compose(
[
# T.RandomHorizontalFlip(p=self.aug_prob),
# T.RandomVerticalFlip(p=self.aug_prob),
T.RandomApply(
[
T.RandomErasing(
p=self.aug_prob, scale=(0.02, 0.15), ratio=(0.2, 1.0)
),
T.RandomAffine(
degrees=(0, 70),
translate=(0.1, 0.2),
scale=(0.95, 1.2),
interpolation=T.InterpolationMode.BILINEAR,
),
],
p=self.aug_prob,
),
# T.RandomAutocontrast(p=self.aug_prob),
# T.RandomAdjustSharpness(sharpness_factor=2, p=self.aug_prob)
# T.RandomErasing(p=self.aug_prob)
]
)
if aug_prob != 0.0
else lambda *x: x
)
def _split_parts(self, file, load_all=True, rgb_to_bgr=False, keys=None, full=False):
# has already been normalized
# warning: key RGB is HRMSI when the dataset is GF5-GF1
if not full:
keys = ['GT', 'LRHSI', 'RGB', 'HSI_up']
else:
keys = ['LRHSI', 'RGB', 'HSI_up']
if load_all:
# load all data in memory
data = []
for k in keys:
data.append(
self.dataset_fn(torch.tensor(file[k][:], dtype=torch.float32)),
)
if rgb_to_bgr:
print("warning: rgb to bgr, for testing generalization only.")
# rgb -> bgr
if not full:
data[2] = data[2][:, [-1, 1, 0]]
else:
data[1] = data[1][:, [-1, 1, 0]]
return data
else:
# warning: it will ignore @normalize
# warning: "GT" can not be access in FULL mode
return (
file.get("GT"),
file.get("LRHSI"),
file.get("RGB"),
file.get("HSI_up"),
)
def aug_trans(self, *data):
data_list = []
seed = torch.random.seed()
for d in data:
torch.manual_seed(seed)
random.seed(seed)
d = self.geo_trans(d)
data_list.append(d)
return tuple(data_list)
def __getitem__(self, index):
# gt: [31, 64, 64]
# lr_hsi: [31, 16, 16]
# rbg: [3, 64, 64]
# hsi_up: [31, 64, 64]
# harvard [rgb]
# cave [bgr]
if not self.full_res:
tuple_data = (
self.rgb[index],
self.lr_hsi[index],
self.hsi_up[index],
self.gt[index],
)
else:
tuple_data = (
self.rgb[index],
self.lr_hsi[index],
self.hsi_up[index],
)
if self.aug_prob != 0.0:
return self.aug_trans(*tuple_data)
else:
return tuple_data
def __len__(self):
return len(self.rgb)
|
294coder/Efficient-MIF | 6,189 | task_datasets/WV3.py | import torch
import torch.utils.data as data
import torchvision.transforms as T
import cv2
import numpy as np
import h5py
from typing import List, Tuple, Optional, Union
class Identity:
def __call__(self, *args):
return args
class WV3Datasets(data.Dataset):
def __init__(
self,
file: Union[h5py.File, str, dict],
aug_prob=0.0,
hp=False,
hp_ksize=(5, 5),
norm_range=True,
full_res=False,
):
"""
:param d: h5py.File or dict
:param aug_prob: augmentation probability
:param hp: high pass for ms and pan. x = x - cv2.boxFilter(x)
:param hp_ksize: cv2.boxFiler kernel size
:param norm_range: normalize data range
"""
super(WV3Datasets, self).__init__()
# FIXME: should pass @path rather than @file which is h5py.File object to avoid can not be pickled error
if isinstance(file, (str, h5py.File)):
if isinstance(file, str):
file = h5py.File(file)
print(
"warning: when @file is a h5py.File object, it can not be pickled.",
"try to set DataLoader number_worker to 0",
)
if not full_res:
self.gt, self.ms, self.lms, self.pan = self.get_divided(file)
print("datasets shape:")
print("{:^20}{:^20}{:^20}{:^20}".format("pan", "ms", "lms", "gt"))
print(
"{:^20}{:^20}{:^20}{:^20}".format(
str(tuple(self.pan.shape)),
str(tuple(self.ms.shape)),
str(tuple(self.lms.shape)),
str(tuple(self.gt.shape)),
)
)
else:
self.ms, self.lms, self.pan = self.get_divided(file, True)
print("datasets shape:")
print("{:^20}{:^20}{:^20}".format("pan", "ms", "lms"))
print(
"{:^20}{:^20}{:^20}".format(
str(tuple(self.pan.shape)),
str(tuple(self.ms.shape)),
str(tuple(self.lms.shape)),
)
)
self.size = self.ms.shape[0]
# highpass filter
self.hp = hp
self.hp_ksize = hp_ksize
if hp and hp_ksize is not None:
self.group_high_pass(hp_ksize)
# to tensor
if norm_range:
def norm_func(x):
x = x / 2047.0
return x
else:
def norm_func(x):
return x
self.pan = norm_func(self.pan)
self.ms = norm_func(self.ms)
self.lms = norm_func(self.lms)
if not full_res:
self.gt = norm_func(self.gt)
# geometrical transformation
self.aug_prob = aug_prob
self.geo_trans = (
T.Compose(
[T.RandomVerticalFlip(p=aug_prob), T.RandomHorizontalFlip(p=aug_prob)]
)
if aug_prob != 0.0
else Identity()
)
@staticmethod
def get_divided(d, full_resolution=False):
if not full_resolution:
return (
torch.tensor(d["gt"][:], dtype=torch.float32),
torch.tensor(d["ms"][:], dtype=torch.float32),
torch.tensor(d["lms"][:], dtype=torch.float32),
torch.tensor(d["pan"][:], dtype=torch.float32),
)
else:
return (
torch.tensor(d["ms"][:], dtype=torch.float32),
torch.tensor(d["lms"][:], dtype=torch.float32),
torch.tensor(d["pan"][:], dtype=torch.float32),
)
@staticmethod
def _get_high_pass(data, k_size):
for i, img in enumerate(data):
hp = cv2.boxFilter(img.transpose(1, 2, 0), -1, k_size)
if hp.ndim == 2:
hp = hp[..., np.newaxis]
data[i] = img - hp.transpose(2, 0, 1)
return data
def group_high_pass(self, k_size):
self.ms = self._get_high_pass(self.ms, k_size)
self.pan = self._get_high_pass(self.pan, k_size)
def aug_trans(self, *data):
data_list = []
seed = torch.random.seed()
for d in data:
torch.manual_seed(seed)
d = self.geo_trans(d)
data_list.append(d)
return data_list
def __getitem__(self, item):
if hasattr(self, "gt"):
tuple_data = (self.pan[item], self.ms[item], self.lms[item], self.gt[item])
else:
tuple_data = (self.pan[item], self.ms[item], self.lms[item])
return self.aug_trans(*tuple_data) if self.aug_prob != 0.0 else tuple_data
def __len__(self):
return self.size
def __repr__(self):
return (
f"num: {self.size} \n "
f"augmentation: {self.geo_trans} \n"
f"get high pass ms and pan: {self.hp} \n "
f"filter kernel size: {self.hp_ksize}"
)
def make_datasets(
path, split_ratio=0.8, hp=True, seed=2022, aug_probs: Tuple = (0.0, 0.0)
):
"""
if your dataset didn't split before, use this function will split your dataset into two part,
which are train and validate datasets.
:param device: device
:param path: datasets path
:param split_ratio: train validate split ratio
:param hp: get high pass data, only works for ms and pan data
:param seed: split data random state
:param aug_probs: augmentation probabilities, type List
:return: List[datasets]
"""
d = h5py.File(path)
ds = [
torch.tensor(d["gt"]),
torch.tensor(d["ms"]),
torch.tensor(d["lms"]),
torch.tensor(d["pan"]),
]
n = ds[0].shape[0]
s = int(n * split_ratio)
random_perm = np.arange(n)
np.random.seed(seed)
np.random.shuffle(random_perm)
train_set = {}
val_set = {}
for i, name in enumerate(["gt", "ms", "lms", "pan"]):
ds[i] = ds[i][random_perm]
train_set[name] = ds[i][:s]
val_set[name] = ds[i][s:]
train_ds = WV3Datasets(train_set, hp=hp, aug_prob=aug_probs[0])
val_ds = WV3Datasets(val_set, hp=hp, aug_prob=aug_probs[1])
return train_ds, val_ds
|
294coder/Efficient-MIF | 8,117 | utils/optim_utils.py | from typing import Iterable, Optional, Union
import weakref
import numpy as np
import torch
import torch.nn as nn
import torch.optim as optim
from torch.optim.lr_scheduler import (
CosineAnnealingLR,
MultiStepLR,
CosineAnnealingWarmRestarts,
ReduceLROnPlateau,
)
import accelerate
import torch_ema
import deepspeed
from deepspeed.runtime.zero import GatheredParameters
import sys
sys.path.append('./')
from utils.misc import is_main_process
class IdentityScheduler(torch.optim.lr_scheduler._LRScheduler):
# a placeholder for lr_scheduler or weight_decay_scheduler
def __init__(self, optim, **kwargs):
self.optim = optim
self.kwargs = kwargs
def step(self, *args, **kwargs):
pass
def state_dict(self):
return self.kwargs
def load_state_dict(self, state_dict):
pass
def cosine_scheduler(
base_value, final_value, epochs, niter_per_ep, warmup_epochs=0, start_warmup_value=0
):
"""
copy from DINO. manually set learning lr every iteration.
note that there is only half epoch of cosine, which means learning rate will not
go back to the original.
:param base_value:
:param final_value:
:param epochs:
:param niter_per_ep:
:param warmup_epochs:
:param start_warmup_value:
:return:
"""
warmup_schedule = np.array([])
warmup_iters = warmup_epochs * niter_per_ep
if warmup_epochs > 0:
warmup_schedule = np.linspace(start_warmup_value, base_value, warmup_iters)
iters = np.arange(epochs * niter_per_ep - warmup_iters)
schedule = final_value + 0.5 * (base_value - final_value) * (
1 + np.cos(np.pi * iters / len(iters))
)
schedule = np.concatenate((warmup_schedule, schedule))
assert len(schedule) == epochs * niter_per_ep
return schedule
class LinearWarmupScheduler:
def __init__(self, opt: optim.Optimizer, init_value, warmup_value, warmup_epochs):
self.opt = opt
self.init_value = init_value
self.warmup_value = warmup_value
self.warmup_epochs = warmup_epochs
self.values = np.linspace(init_value, warmup_value, warmup_epochs)
self.now_index = 0
def step(self):
self.opt.param_groups[0]["lr"] = self.values[self.now_index]
self.now_index += 1
class CosineAnnealingWarmRestartsReduce(CosineAnnealingWarmRestarts):
def __init__(self, opt: optim.Optimizer, T_0, T_mult=1, lr_mult=1, eta_min=0, last_epoch=-1):
self.opt = opt
self.lr_mult = lr_mult
super().__init__(opt, T_0, T_mult, eta_min, last_epoch)
def step(self, epoch=None):
super().step(epoch)
if self.T_cur == self.T_i-1 and self.last_epoch != 0:
# reduce the base lr
for i in range(len(self.base_lrs)):
self.base_lrs[i] *= self.lr_mult
self.base_lrs[i] = max(self.base_lrs[i], self.eta_min)
def get_precision(mixed_precision):
if mixed_precision == 'fp32' or mixed_precision == 'no':
return torch.float32
elif mixed_precision == 'fp16':
return torch.float16
elif mixed_precision == 'bf16':
return torch.bfloat16
else:
raise ValueError(f"Invalid mixed precision value: {mixed_precision}")
def get_scheduler(optim, **kwargs):
"""
get lr_scheduler or weight_decay_scheduler
Args:
optim: optimizer
**kwargs: a dict containing type of scheduler and its arguments
Returns: a scheduler
"""
name = kwargs["name"]
kwargs.pop("name")
if name == "cos_anneal":
return CosineAnnealingLR(optim, **kwargs)
elif name == "cos_anneal_restart":
return CosineAnnealingWarmRestarts(optim, **kwargs)
elif name == "cos_anneal_restart_reduce":
return CosineAnnealingWarmRestartsReduce(optim, **kwargs)
elif name == "multi_step":
return MultiStepLR(optim, **kwargs)
elif name == "plateau":
return ReduceLROnPlateau(optim, **kwargs)
elif name == "identity":
return IdentityScheduler(optim, **kwargs)
else:
raise NotImplementedError
def get_optimizer(model: torch.nn.Module, params: "Iterable | dict", **kwargs):
name = kwargs["name"]
kwargs.pop("name")
if name == "sgd":
return optim.SGD(params, **kwargs)
elif name == "adam":
return optim.Adam(params, **kwargs)
elif name == "adamw":
return optim.AdamW(params, **kwargs)
elif name == 'lion':
from lion_pytorch import Lion
return Lion(params, betas=(0.95, 0.98), use_triton=True, **kwargs)
elif name == 'fusedadam':
return deepspeed.ops.adam.FusedAdam(params, **kwargs)
elif name == 'schedulefree-adam':
import schedulefree
return schedulefree.AdamWScheduleFree(params, **kwargs)
elif name == 'adam-mini':
return Adam_mini(model, **kwargs)
else:
raise NotImplementedError(f'optimizer {name} not implemented')
def get_ema_model(parameters: Iterable[torch.nn.Parameter],
accelerator: accelerate.Accelerator=None,
**ema_kwargs):
if accelerator is not None:
if accelerator.state.deepspeed_plugin is not None:
ema_model = DeepspeedEMA(parameters, **ema_kwargs)
return ema_model
ema_model = torch_ema.ema.ExponentialMovingAverage(model, **ema_kwargs)
return ema_model
if __name__ == "__main__":
import matplotlib.pyplot as plt
import torch.optim as optim
import torch.nn as nn
# init_lr = 1e-3
# final_lr = 1e-6
# epochs = 500
# # nither_per_ep = int(np.ceil(3000 // 16)) # len(datasets) / batch_size
# # warm_epochs = 80
# # start_warmup_value = init_lr
# # cos_sche = cosine_scheduler(
# # init_lr, final_lr, epochs, nither_per_ep, warm_epochs, start_warmup_value
# # )
# # plt.plot(list(map(lambda x: x / nither_per_ep, range(len(cos_sche)))), cos_sche)
# # plt.show()
# # torch cosine annealing lr scheduler
# net = nn.Sequential(nn.Linear(8, 64))
# optimizer = optim.AdamW(net.parameters(), lr=init_lr)
# # cos_sche2 = CosineAnnealingLR(optimizer, epochs - warm_epochs, final_lr)
# cos_anneal_reduce_sche = CosineAnnealingWarmRestartsReduce(optimizer, 50, 2, 0.5, 1e-6, last_epoch=-1)
# lr = []
# for i in range(200, 500):
# l = optimizer.param_groups[0]["lr"]
# lr.append(l)
# # if i > warm_epochs:
# # cos_sche2.step()
# cos_anneal_reduce_sche.step(i)
# plt.plot(range(200, 500), lr)
# # plt.show()
# plt.savefig('cos_anneal_reduce.png')
import accelerate
from torch.utils.data import DataLoader, Dataset, TensorDataset
accelerator = accelerate.Accelerator()
model = torch.nn.Linear(3, 16)
opt = optim.Adam(model.parameters(), lr=1e-3)
dl = DataLoader(TensorDataset(*([torch.randn(16, 3)]*2)), batch_size=1)
model, dl, opt = accelerator.prepare(model, dl, opt)
ema_model = get_ema_model(list(model.parameters()), accelerator=accelerator, decay=0.9)
accelerator.wait_for_everyone()
if is_main_process():
print('main process: set weight to zero weight \n')
model.weight.data.zero_()
model.bias.data.zero_()
print(model.weight)
else:
import time
time.sleep(2)
# the other process
print('other process: set weight to non-zero weight\n')
print(ema_model.shadow_params)
accelerator.wait_for_everyone()
print('-----------------------------'*2, '\n')
ema_model.update()
# proc 0 is zeros but proc 1 is not,
# so the ema_model should not be all zeros
if is_main_process():
print('main process')
print(ema_model.shadow_params)
print(model.weight)
print('-----------------------------'*2, '\n')
ema_model.restore(model.parameters())
print(model.weight)
print('-----------------------------'*2)
|
2977094657/DidaAPI | 4,920 | frontend/docs/.vitepress/theme/style/vars.css | /**
* Customize default theme styling by overriding CSS variables:
* https://github.com/vuejs/vitepress/blob/main/src/client/theme-default/styles/vars.css
*/
/**
* Colors
*
* Each colors have exact same color scale system with 3 levels of solid
* colors with different brightness, and 1 soft color.
*
* - `XXX-1`: The most solid color used mainly for colored text. It must
* satisfy the contrast ratio against when used on top of `XXX-soft`.
*
* - `XXX-2`: The color used mainly for hover state of the button.
*
* - `XXX-3`: The color for solid background, such as bg color of the button.
* It must satisfy the contrast ratio with pure white (#ffffff) text on
* top of it.
*
* - `XXX-soft`: The color used for subtle background such as custom container
* or badges. It must satisfy the contrast ratio when putting `XXX-1` colors
* on top of it.
*
* The soft color must be semi transparent alpha channel. This is crucial
* because it allows adding multiple "soft" colors on top of each other
* to create a accent, such as when having inline code block inside
* custom containers.
*
* - `default`: The color used purely for subtle indication without any
* special meanings attched to it such as bg color for menu hover state.
*
* - `brand`: Used for primary brand colors, such as link text, button with
* brand theme, etc.
*
* - `tip`: Used to indicate useful information. The default theme uses the
* brand color for this by default.
*
* - `warning`: Used to indicate warning to the users. Used in custom
* container, badges, etc.
*
* - `danger`: Used to indicate dangerous message to the users. Used in custom
* container, badges, etc.
* -------------------------------------------------------------------------- */
:root {
--vp-c-default-1: var(--vp-c-gray-1);
--vp-c-default-2: var(--vp-c-gray-2);
--vp-c-default-3: var(--vp-c-gray-3);
--vp-c-default-soft: var(--vp-c-gray-soft);
/* 使用滴答清单官方色彩 */
--vp-c-brand-1: #4772FA;
--vp-c-brand-2: #5a7efb;
--vp-c-brand-3: #6d8afc;
--vp-c-brand-soft: rgba(71, 114, 250, 0.14);
--vp-c-tip-1: var(--vp-c-brand-1);
--vp-c-tip-2: var(--vp-c-brand-2);
--vp-c-tip-3: var(--vp-c-brand-3);
--vp-c-tip-soft: var(--vp-c-brand-soft);
/* 使用滴答清单的橙色作为警告色 */
--vp-c-warning-1: #FFB000;
--vp-c-warning-2: #ffba1a;
--vp-c-warning-3: #ffc433;
--vp-c-warning-soft: rgba(255, 176, 0, 0.14);
--vp-c-danger-1: var(--vp-c-red-1);
--vp-c-danger-2: var(--vp-c-red-2);
--vp-c-danger-3: var(--vp-c-red-3);
--vp-c-danger-soft: var(--vp-c-red-soft);
}
/**
* Component: Button
* -------------------------------------------------------------------------- */
:root {
--vp-button-brand-border: transparent;
--vp-button-brand-text: var(--vp-c-white);
--vp-button-brand-bg: var(--vp-c-brand-3);
--vp-button-brand-hover-border: transparent;
--vp-button-brand-hover-text: var(--vp-c-white);
--vp-button-brand-hover-bg: var(--vp-c-brand-2);
--vp-button-brand-active-border: transparent;
--vp-button-brand-active-text: var(--vp-c-white);
--vp-button-brand-active-bg: var(--vp-c-brand-1);
}
/**
* Component: Home
* -------------------------------------------------------------------------- */
:root {
--vp-home-hero-name-color: transparent;
--vp-home-hero-name-background: -webkit-linear-gradient(
120deg,
#4772FA 30%,
#FFB000
);
--vp-home-hero-image-background-image: linear-gradient(
-45deg,
#4772FA 50%,
#FFB000 50%
);
--vp-home-hero-image-filter: blur(44px);
}
@media (min-width: 640px) {
:root {
--vp-home-hero-image-filter: blur(56px);
}
}
@media (min-width: 960px) {
:root {
--vp-home-hero-image-filter: blur(68px);
}
}
/**
* Component: Custom Block
* -------------------------------------------------------------------------- */
:root {
--vp-custom-block-tip-border: transparent;
--vp-custom-block-tip-text: var(--vp-c-text-1);
--vp-custom-block-tip-bg: var(--vp-c-brand-soft);
--vp-custom-block-tip-code-bg: var(--vp-c-brand-soft);
}
/**
* Component: Algolia
* -------------------------------------------------------------------------- */
.DocSearch {
--docsearch-primary-color: var(--vp-c-brand-1) !important;
}
/**
* Component: Layout adjustments
* -------------------------------------------------------------------------- */
/* 增加首页顶部距离 */
.VPHome {
padding-top: 40px !important;
}
/* 调整Hero区域间距 */
.VPHero .container {
padding-top: 30px !important;
padding-bottom: 40px !important;
}
/* 调整Features区域间距 */
.VPFeatures {
padding-top: 40px !important;
padding-bottom: 80px !important;
}
/* 响应式调整 */
@media (max-width: 768px) {
.VPHome {
padding-top: 40px !important;
}
.VPHero .container {
padding-top: 40px !important;
padding-bottom: 40px !important;
}
.VPFeatures {
padding-top: 20px !important;
padding-bottom: 40px !important;
}
}
|
2977094657/DidaAPI | 2,468 | frontend/docs/api/habits/week-current-statistics.md | # 获取本周习惯打卡统计
## 接口信息
- **接口路径**: `GET /habits/statistics/week/current`
- **接口描述**: 获取本周的习惯打卡统计信息
- **请求方式**: GET
- **认证要求**: 需要先完成微信登录获取认证会话
## 请求参数
无需参数
## 请求示例
```bash
curl -X GET "http://localhost:8000/habits/statistics/week/current"
```
## 响应格式
### 成功响应
```json
{
"weekStart": "2023-12-04",
"weekEnd": "2023-12-10",
"totalHabits": 5,
"completedDays": 4,
"completionRate": 80.0,
"dailyStats": [
{
"date": "2023-12-04",
"dayOfWeek": "Monday",
"completedHabits": 4,
"totalHabits": 5,
"completionRate": 80.0
},
{
"date": "2023-12-05",
"dayOfWeek": "Tuesday",
"completedHabits": 5,
"totalHabits": 5,
"completionRate": 100.0
}
],
"habitDetails": [
{
"habitId": "habit123",
"habitName": "早起",
"targetDays": 7,
"completedDays": 5,
"completionRate": 71.4,
"streak": 3,
"dailyStatus": [
{
"date": "2023-12-04",
"completed": true
},
{
"date": "2023-12-05",
"completed": true
}
]
}
]
}
```
### 错误响应
#### 未认证
```json
{
"error": "no_auth_session",
"message": "未设置认证会话,请先完成微信登录"
}
```
#### 服务错误
```json
{
"error": "service_error",
"message": "获取本周习惯打卡统计失败,请稍后重试"
}
```
## 响应字段说明
| 字段名 | 类型 | 描述 |
|--------|------|------|
| weekStart | string | 本周开始日期 |
| weekEnd | string | 本周结束日期 |
| totalHabits | integer | 总习惯数量 |
| completedDays | integer | 完成打卡的天数 |
| completionRate | float | 整体完成率(百分比) |
| dailyStats | array | 每日统计数据 |
| dailyStats[].date | string | 日期 |
| dailyStats[].dayOfWeek | string | 星期几 |
| dailyStats[].completedHabits | integer | 当日完成的习惯数 |
| dailyStats[].totalHabits | integer | 当日总习惯数 |
| dailyStats[].completionRate | float | 当日完成率 |
| habitDetails | array | 习惯详细信息 |
| habitDetails[].habitId | string | 习惯ID |
| habitDetails[].habitName | string | 习惯名称 |
| habitDetails[].targetDays | integer | 目标天数 |
| habitDetails[].completedDays | integer | 已完成天数 |
| habitDetails[].completionRate | float | 习惯完成率 |
| habitDetails[].streak | integer | 连续打卡天数 |
| habitDetails[].dailyStatus | array | 每日打卡状态 |
## 使用说明
1. **认证要求**: 需要先调用微信登录接口获取认证会话
2. **统计周期**: 统计当前自然周(周一到周日)的数据
3. **实时更新**: 数据实时更新,反映最新的打卡状态
4. **多维度统计**: 提供整体、每日、每个习惯的多维度统计
## 相关接口
- [获取所有习惯](../habits.md)
- [导出习惯数据](./export-habits.md)
- [微信登录流程](../auth/wechat-login-flow.md)
## 注意事项
- 需要先完成微信登录获取认证会话
- 统计数据基于当前自然周
- 完成率保留一位小数
- 连续打卡天数会在中断后重新计算
|
281677160/openwrt-package | 72,103 | luci-app-passwall/root/usr/share/passwall/iptables.sh | #!/bin/sh
DIR="$(cd "$(dirname "$0")" && pwd)"
MY_PATH=$DIR/iptables.sh
IPSET_LOCAL="passwall_local"
IPSET_LAN="passwall_lan"
IPSET_VPS="passwall_vps"
IPSET_SHUNT="passwall_shunt"
IPSET_GFW="passwall_gfw"
IPSET_CHN="passwall_chn"
IPSET_BLACK="passwall_black"
IPSET_WHITE="passwall_white"
IPSET_BLOCK="passwall_block"
IPSET_LOCAL6="passwall_local6"
IPSET_LAN6="passwall_lan6"
IPSET_VPS6="passwall_vps6"
IPSET_SHUNT6="passwall_shunt6"
IPSET_GFW6="passwall_gfw6"
IPSET_CHN6="passwall_chn6"
IPSET_BLACK6="passwall_black6"
IPSET_WHITE6="passwall_white6"
IPSET_BLOCK6="passwall_block6"
FORCE_INDEX=2
USE_SHUNT_TCP=0
USE_SHUNT_UDP=0
. /lib/functions/network.sh
ipt=$(command -v iptables-legacy || command -v iptables)
ip6t=$(command -v ip6tables-legacy || command -v ip6tables)
ipt_n="$ipt -t nat -w"
ipt_m="$ipt -t mangle -w"
ip6t_n="$ip6t -t nat -w"
ip6t_m="$ip6t -t mangle -w"
[ -z "$ip6t" -o -z "$(lsmod | grep 'ip6table_nat')" ] && ip6t_n="eval #$ip6t_n"
[ -z "$ip6t" -o -z "$(lsmod | grep 'ip6table_mangle')" ] && ip6t_m="eval #$ip6t_m"
FWI=$(uci -q get firewall.passwall.path 2>/dev/null)
FAKE_IP="198.18.0.0/15"
FAKE_IP_6="fc00::/18"
factor() {
if [ -z "$1" ] || [ -z "$2" ]; then
echo ""
elif [ "$1" == "1:65535" ]; then
echo ""
else
echo "$2 $1"
fi
}
dst() {
echo "-m set $2 --match-set $1 dst"
}
comment() {
local name=$(echo $1 | sed 's/ /_/g')
echo "-m comment --comment '$name'"
}
#解决端口超过15个ipt无效,支持单端口、端口范围
add_port_rules() {
local ipt_cmd="$1"
local port_list="$2"
local target="$3"
echo "$port_list" | grep -vq '[0-9]' && return
port_list=$(echo "$port_list" | tr -d ' ' | sed 's/-/:/g' | tr ',' '\n' | awk '!a[$0]++' | grep -v '^$')
[ -z "$port_list" ] && return
if echo "$port_list" | grep -q '^1:65535$'; then
eval "$ipt_cmd $target"
return
fi
local multiport_ports=""
local range_ports=""
local count=0
local port
for port in $port_list; do
if echo "$port" | grep -q ':'; then
range_ports="$range_ports $port"
else
multiport_ports="$multiport_ports,$port"
count=$((count + 1))
if [ "$count" -eq 15 ]; then
eval "$ipt_cmd -m multiport --dport ${multiport_ports#,} $target"
multiport_ports=""
count=0
fi
fi
done
if [ -n "$multiport_ports" ]; then
eval "$ipt_cmd -m multiport --dport ${multiport_ports#,} $target"
fi
for port in $range_ports; do
eval "$ipt_cmd --dport $port $target"
done
}
destroy_ipset() {
for i in "$@"; do
ipset -q -F $i
ipset -q -X $i
done
}
insert_rule_before() {
[ $# -ge 3 ] || {
return 1
}
local ipt_tmp="${1}"; shift
local chain="${1}"; shift
local keyword="${1}"; shift
local rule="${1}"; shift
local default_index="${1}"; shift
default_index=${default_index:-0}
local _index=$($ipt_tmp -n -L $chain --line-numbers 2>/dev/null | grep "$keyword" | head -n 1 | awk '{print $1}')
if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then
$ipt_tmp -A $chain $rule
else
if [ -z "${_index}" ]; then
_index=${default_index}
fi
$ipt_tmp -I $chain $_index $rule
fi
}
insert_rule_after() {
[ $# -ge 3 ] || {
return 1
}
local ipt_tmp="${1}"; shift
local chain="${1}"; shift
local keyword="${1}"; shift
local rule="${1}"; shift
local default_index="${1}"; shift
default_index=${default_index:-0}
local _index=$($ipt_tmp -n -L $chain --line-numbers 2>/dev/null | grep "$keyword" | awk 'END {print}' | awk '{print $1}')
if [ -z "${_index}" ] && [ "${default_index}" = "0" ]; then
$ipt_tmp -A $chain $rule
else
if [ -n "${_index}" ]; then
_index=$((_index + 1))
else
_index=${default_index}
fi
$ipt_tmp -I $chain $_index $rule
fi
}
RULE_LAST_INDEX() {
[ $# -ge 3 ] || {
echolog "索引列举方式不正确(iptables),终止执行!"
return 1
}
local ipt_tmp="${1}"; shift
local chain="${1}"; shift
local list="${1}"; shift
local default="${1:-0}"; shift
local _index=$($ipt_tmp -n -L $chain --line-numbers 2>/dev/null | grep "$list" | head -n 1 | awk '{print $1}')
echo "${_index:-${default}}"
}
REDIRECT() {
local s="-j REDIRECT"
if [ -n "$1" ]; then
s="$s --to-ports $1"
if [ "$2" = "MARK" ]; then
s="-j MARK --set-mark $1"
elif [ "$2" = "TPROXY" ]; then
local mark="-m mark --mark 1"
s="${mark} -j TPROXY --tproxy-mark 1/1 --on-port $1"
fi
fi
echo "$s"
}
get_jump_ipt() {
case "$1" in
direct)
local mark="-m mark ! --mark 1"
s="${mark} -j RETURN"
echo $s
;;
proxy)
if [ -n "$2" ] && [ -n "$(echo $2 | grep "^-")" ]; then
echo "$2"
else
echo "$(REDIRECT $2 $3)"
fi
;;
esac
}
gen_lanlist() {
cat $RULES_PATH/lanlist_ipv4 | tr -s '\n' | grep -v "^#"
}
gen_lanlist_6() {
cat $RULES_PATH/lanlist_ipv6 | tr -s '\n' | grep -v "^#"
}
get_wan_ip() {
local NET_IF
local NET_ADDR
network_flush_cache
network_find_wan NET_IF
network_get_ipaddr NET_ADDR "${NET_IF}"
echo $NET_ADDR
}
get_wan6_ip() {
local NET_IF
local NET_ADDR
network_flush_cache
network_find_wan6 NET_IF
network_get_ipaddr6 NET_ADDR "${NET_IF}"
echo $NET_ADDR
}
load_acl() {
([ "$ENABLED_ACLS" == 1 ] || ([ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ])) && echolog " - 访问控制:"
[ "$ENABLED_ACLS" == 1 ] && {
acl_app
for sid in $(ls -F ${TMP_ACL_PATH} | grep '/$' | awk -F '/' '{print $1}' | grep -v 'default'); do
eval $(uci -q show "${CONFIG}.${sid}" | cut -d'.' -sf 3-)
tcp_no_redir_ports=${tcp_no_redir_ports:-default}
udp_no_redir_ports=${udp_no_redir_ports:-default}
use_global_config=${use_global_config:-0}
tcp_proxy_drop_ports=${tcp_proxy_drop_ports:-default}
udp_proxy_drop_ports=${udp_proxy_drop_ports:-default}
tcp_redir_ports=${tcp_redir_ports:-default}
udp_redir_ports=${udp_redir_ports:-default}
use_direct_list=${use_direct_list:-1}
use_proxy_list=${use_proxy_list:-1}
use_block_list=${use_block_list:-1}
use_gfw_list=${use_gfw_list:-1}
chn_list=${chn_list:-direct}
tcp_proxy_mode=${tcp_proxy_mode:-proxy}
udp_proxy_mode=${udp_proxy_mode:-proxy}
[ "$tcp_no_redir_ports" = "default" ] && tcp_no_redir_ports=$TCP_NO_REDIR_PORTS
[ "$udp_no_redir_ports" = "default" ] && udp_no_redir_ports=$UDP_NO_REDIR_PORTS
[ "$tcp_proxy_drop_ports" = "default" ] && tcp_proxy_drop_ports=$TCP_PROXY_DROP_PORTS
[ "$udp_proxy_drop_ports" = "default" ] && udp_proxy_drop_ports=$UDP_PROXY_DROP_PORTS
[ "$tcp_redir_ports" = "default" ] && tcp_redir_ports=$TCP_REDIR_PORTS
[ "$udp_redir_ports" = "default" ] && udp_redir_ports=$UDP_REDIR_PORTS
[ -n "$(get_cache_var "ACL_${sid}_tcp_node")" ] && tcp_node=$(get_cache_var "ACL_${sid}_tcp_node")
[ -n "$(get_cache_var "ACL_${sid}_tcp_redir_port")" ] && tcp_port=$(get_cache_var "ACL_${sid}_tcp_redir_port")
[ -n "$(get_cache_var "ACL_${sid}_udp_node")" ] && udp_node=$(get_cache_var "ACL_${sid}_udp_node")
[ -n "$(get_cache_var "ACL_${sid}_udp_redir_port")" ] && udp_port=$(get_cache_var "ACL_${sid}_udp_redir_port")
[ -n "$(get_cache_var "ACL_${sid}_dns_port")" ] && dns_redirect_port=$(get_cache_var "ACL_${sid}_dns_port")
[ -n "$tcp_node" ] && tcp_node_remark=$(config_n_get $tcp_node remarks)
[ -n "$udp_node" ] && udp_node_remark=$(config_n_get $udp_node remarks)
use_shunt_tcp=0
use_shunt_udp=0
[ -n "$tcp_node" ] && [ "$(config_n_get $tcp_node protocol)" = "_shunt" ] && use_shunt_tcp=1
[ -n "$udp_node" ] && [ "$(config_n_get $udp_node protocol)" = "_shunt" ] && use_shunt_udp=1
[ "${use_global_config}" = "1" ] && {
tcp_node_remark=$(config_n_get $TCP_NODE remarks)
udp_node_remark=$(config_n_get $UDP_NODE remarks)
use_direct_list=${USE_DIRECT_LIST}
use_proxy_list=${USE_PROXY_LIST}
use_block_list=${USE_BLOCK_LIST}
use_gfw_list=${USE_GFW_LIST}
chn_list=${CHN_LIST}
tcp_proxy_mode=${TCP_PROXY_MODE}
udp_proxy_mode=${UDP_PROXY_MODE}
use_shunt_tcp=${USE_SHUNT_TCP}
use_shunt_udp=${USE_SHUNT_UDP}
dns_redirect_port=${DNS_REDIRECT_PORT}
black_set_name=${IPSET_BLACK}
black6_set_name=${IPSET_BLACK6}
gfw_set_name=${IPSET_GFW}
gfw6_set_name=${IPSET_GFW6}
shunt_set_name=${IPSET_SHUNT}
shunt6_set_name=${IPSET_SHUNT6}
}
_acl_list=${TMP_ACL_PATH}/${sid}/source_list
for i in $(cat $_acl_list); do
local _ipt_source _ipv4
local msg
if [ -n "${interface}" ]; then
. /lib/functions/network.sh
local gateway device
network_get_gateway gateway "${interface}"
network_get_device device "${interface}"
[ -z "${device}" ] && device="${interface}"
_ipt_source="-i ${device} "
msg="源接口【${device}】,"
else
msg="源接口【所有】,"
fi
if [ -n "$(echo ${i} | grep '^iprange:')" ]; then
_iprange=$(echo ${i} | sed 's#iprange:##g')
_ipt_source=$(factor ${_iprange} "${_ipt_source}-m iprange --src-range")
msg="${msg}IP range【${_iprange}】,"
_ipv4="1"
unset _iprange
elif [ -n "$(echo ${i} | grep '^ipset:')" ]; then
_ipset=$(echo ${i} | sed 's#ipset:##g')
msg="${msg}IPset【${_ipset}】,"
ipset -q list ${_ipset} >/dev/null
if [ $? -eq 0 ]; then
_ipt_source="${_ipt_source}-m set --match-set ${_ipset} src"
unset _ipset
else
echolog " - 【$remarks】,${msg}不存在,忽略。"
unset _ipset
continue
fi
elif [ -n "$(echo ${i} | grep '^ip:')" ]; then
_ip=$(echo ${i} | sed 's#ip:##g')
_ipt_source=$(factor ${_ip} "${_ipt_source}-s")
msg="${msg}IP【${_ip}】,"
_ipv4="1"
unset _ip
elif [ -n "$(echo ${i} | grep '^mac:')" ]; then
_mac=$(echo ${i} | sed 's#mac:##g')
_ipt_source=$(factor ${_mac} "${_ipt_source}-m mac --mac-source")
msg="${msg}MAC【${_mac}】,"
unset _mac
elif [ -n "$(echo ${i} | grep '^any')" ]; then
msg="${msg}所有设备,"
else
continue
fi
msg="【$remarks】,${msg}"
ipt_tmp=$ipt_n
[ -n "${is_tproxy}" ] && ipt_tmp=$ipt_m
[ "$tcp_no_redir_ports" != "disable" ] && {
if ! has_1_65535 "$tcp_no_redir_ports"; then
[ "$_ipv4" != "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p tcp" $tcp_no_redir_ports "-j RETURN" 2>/dev/null
add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") ${_ipt_source} -p tcp" $tcp_no_redir_ports "-j RETURN"
echolog " - ${msg}不代理 TCP 端口[${tcp_no_redir_ports}]"
else
#结束时会return,无需加多余的规则。
unset tcp_port
echolog " - ${msg}不代理所有 TCP 端口"
fi
}
[ "$udp_no_redir_ports" != "disable" ] && {
if ! has_1_65535 "$udp_no_redir_ports"; then
[ "$_ipv4" != "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp" $udp_no_redir_ports "-j RETURN" 2>/dev/null
add_port_rules "$ipt_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp" $udp_no_redir_ports "-j RETURN"
echolog " - ${msg}不代理 UDP 端口[${udp_no_redir_ports}]"
else
#结束时会return,无需加多余的规则。
unset udp_port
echolog " - ${msg}不代理所有 UDP 端口"
fi
}
local dns_redirect
[ $(config_t_get global dns_redirect "1") = "1" ] && dns_redirect=53
if ([ -n "$tcp_port" ] && [ -n "${tcp_proxy_mode}" ]) || ([ -n "$udp_port" ] && [ -n "${udp_proxy_mode}" ]); then
[ "${use_proxy_list}" = "1" ] && {
[ "${use_global_config}" = "0" ] && {
black_set_name="passwall_${sid}_black"
black6_set_name="passwall_${sid}_black6"
ipset -! create $black_set_name nethash maxelem 1048576 timeout 172800
ipset -! create $black6_set_name nethash family inet6 maxelem 1048576 timeout 172800
}
}
[ "${use_gfw_list}" = "1" ] && {
[ "${use_global_config}" = "0" ] && {
gfw_set_name="passwall_${sid}_gfw"
gfw6_set_name="passwall_${sid}_gfw6"
ipset -! create $gfw_set_name nethash maxelem 1048576 timeout 172800
ipset -! create $gfw6_set_name nethash family inet6 maxelem 1048576 timeout 172800
}
}
[ "${use_shunt_tcp}" = "1" -o "${use_shunt_udp}" = "1" ] && {
[ "${use_global_config}" = "0" ] && {
shunt_set_name="passwall_${sid}_shunt"
shunt6_set_name="passwall_${sid}_shunt6"
ipset -! create $shunt_set_name nethash maxelem 1048576 timeout 172800
ipset -! create $shunt6_set_name nethash family inet6 maxelem 1048576 timeout 172800
}
}
[ -n "${dns_redirect_port}" ] && dns_redirect=${dns_redirect_port}
else
[ -n "${DIRECT_DNSMASQ_PORT}" ] && dns_redirect=${DIRECT_DNSMASQ_PORT}
fi
if [ -n "${dns_redirect}" ]; then
$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j RETURN
[ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j RETURN 2>/dev/null
$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j RETURN
[ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j RETURN 2>/dev/null
$ipt_n -A PSW_DNS $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect}
[ "$_ipv4" != "1" ] && $ip6t_n -A PSW_DNS $(comment "$remarks") -p udp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect} 2>/dev/null
$ipt_n -A PSW_DNS $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect}
[ "$_ipv4" != "1" ] && $ip6t_n -A PSW_DNS $(comment "$remarks") -p tcp ${_ipt_source} --dport 53 -j REDIRECT --to-ports ${dns_redirect} 2>/dev/null
[ -z "$(get_cache_var "ACL_${sid}_tcp_default")" ] && echolog " - ${msg}使用与全局配置不相同节点,已将DNS强制重定向到专用 DNS 服务器。"
fi
[ -n "$tcp_port" -o -n "$udp_port" ] && {
[ "${use_block_list}" = "1" ] && $ipt_m -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_BLOCK) -j DROP
[ "${use_direct_list}" = "1" ] && $ipt_tmp -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_WHITE) -j RETURN
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
[ "${use_block_list}" = "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_BLOCK6) -j DROP 2>/dev/null
[ "${use_direct_list}" = "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} $(dst $IPSET_WHITE6) -j RETURN 2>/dev/null
}
[ "$tcp_proxy_drop_ports" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-d $FAKE_IP_6 -j DROP" 2>/dev/null
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $black6_set_name) -j DROP" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $gfw6_set_name) -j DROP" 2>/dev/null
[ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j DROP")" 2>/dev/null
[ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $shunt6_set_name) -j DROP" 2>/dev/null
[ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-j DROP" 2>/dev/null
}
add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-d $FAKE_IP -j DROP"
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $black_set_name) -j DROP"
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $gfw_set_name) -j DROP"
[ "${chn_list}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "-j DROP")"
[ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "$(dst $shunt_set_name) -j DROP"
[ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_proxy_drop_ports "-j DROP"
echolog " - ${msg}屏蔽代理 TCP 端口[${tcp_proxy_drop_ports}]"
}
[ "$udp_proxy_drop_ports" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-d $FAKE_IP_6 -j DROP" 2>/dev/null
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $black6_set_name) -j DROP" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $gfw6_set_name) -j DROP" 2>/dev/null
[ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j DROP")" 2>/dev/null
[ "${use_shunt_udp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $shunt6_set_name) -j DROP" 2>/dev/null
[ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-j DROP" 2>/dev/null
}
add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-d $FAKE_IP -j DROP"
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $black_set_name) -j DROP"
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $gfw_set_name) -j DROP"
[ "${chn_list}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "-j DROP")"
[ "${use_shunt_udp}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "$(dst $shunt_set_name) -j DROP"
[ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_proxy_drop_ports "-j DROP"
echolog " - ${msg}屏蔽代理 UDP 端口[${udp_proxy_drop_ports}]"
}
}
[ -n "$tcp_port" ] && {
if [ -n "${tcp_proxy_mode}" ]; then
msg2="${msg}使用 TCP 节点[$tcp_node_remark]"
if [ -n "${is_tproxy}" ]; then
msg2="${msg2}(TPROXY:${tcp_port})"
ipt_tmp=$ipt_m
ipt_j="-j PSW_RULE"
else
msg2="${msg2}(REDIRECT:${tcp_port})"
ipt_j="$(REDIRECT $tcp_port)"
fi
[ "$accept_icmp" = "1" ] && {
$ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} -d $FAKE_IP $(REDIRECT)
[ "${use_proxy_list}" = "1" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $black_set_name) $(REDIRECT)
[ "${use_gfw_list}" = "1" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $gfw_set_name) $(REDIRECT)
[ "${chn_list}" != "0" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $IPSET_CHN) $(get_jump_ipt ${chn_list})
[ "${use_shunt_tcp}" = "1" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(dst $shunt_set_name) $(REDIRECT)
[ "${tcp_proxy_mode}" != "disable" ] && $ipt_n -A PSW $(comment "$remarks") -p icmp ${_ipt_source} $(REDIRECT)
}
[ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
$ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} -d $FAKE_IP_6 $(REDIRECT) 2>/dev/null
[ "${use_proxy_list}" = "1" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $black6_set_name) $(REDIRECT) 2>/dev/null
[ "${use_gfw_list}" = "1" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $gfw6_set_name) $(REDIRECT) 2>/dev/null
[ "${chn_list}" != "0" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list}) 2>/dev/null
[ "${use_shunt_tcp}" = "1" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(dst $shunt6_set_name) $(REDIRECT) 2>/dev/null
[ "${tcp_proxy_mode}" != "disable" ] && $ip6t_n -A PSW $(comment "$remarks") -p ipv6-icmp ${_ipt_source} $(REDIRECT) 2>/dev/null
}
$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source} -d $FAKE_IP ${ipt_j}
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $black_set_name) ${ipt_j}"
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $gfw_set_name) ${ipt_j}"
[ "${chn_list}" != "0" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "${ipt_j}")"
[ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $shunt_set_name) ${ipt_j}"
[ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_tmp -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "${ipt_j}"
[ -n "${is_tproxy}" ] && $ipt_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY)
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} -d $FAKE_IP_6 -j PSW_RULE 2>/dev/null
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $black6_set_name) -j PSW_RULE" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $gfw6_set_name) -j PSW_RULE" 2>/dev/null
[ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j PSW_RULE")" 2>/dev/null
[ "${use_shunt_tcp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "$(dst $shunt6_set_name) -j PSW_RULE" 2>/dev/null
[ "${tcp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source}" $tcp_redir_ports "-j PSW_RULE" 2>/dev/null
$ip6t_m -A PSW $(comment "$remarks") -p tcp ${_ipt_source} $(REDIRECT $tcp_port TPROXY) 2>/dev/null
}
else
msg2="${msg}不代理 TCP"
fi
echolog " - ${msg2}"
}
[ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p tcp -j RETURN 2>/dev/null
$ipt_tmp -A PSW $(comment "$remarks") ${_ipt_source} -p tcp -j RETURN
[ -n "$udp_port" ] && {
if [ -n "${udp_proxy_mode}" ]; then
msg2="${msg}使用 UDP 节点[$udp_node_remark]"
msg2="${msg2}(TPROXY:${udp_port})"
$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} -d $FAKE_IP -j PSW_RULE
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $black_set_name) -j PSW_RULE"
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $gfw_set_name) -j PSW_RULE"
[ "${chn_list}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $IPSET_CHN) $(get_jump_ipt ${chn_list} "-j PSW_RULE")"
[ "${use_shunt_udp}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $shunt_set_name) -j PSW_RULE"
[ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "-j PSW_RULE"
$ipt_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} $(REDIRECT $udp_port TPROXY)
[ "$PROXY_IPV6" == "1" ] && [ "$_ipv4" != "1" ] && {
$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} -d $FAKE_IP_6 -j PSW_RULE 2>/dev/null
[ "${use_proxy_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $black6_set_name) -j PSW_RULE" 2>/dev/null
[ "${use_gfw_list}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $gfw6_set_name) -j PSW_RULE" 2>/dev/null
[ "${chn_list}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $IPSET_CHN6) $(get_jump_ipt ${chn_list} "-j PSW_RULE")" 2>/dev/null
[ "${use_shunt_udp}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "$(dst $shunt6_set_name) -j PSW_RULE" 2>/dev/null
[ "${udp_proxy_mode}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source}" $udp_redir_ports "-j PSW_RULE" 2>/dev/null
$ip6t_m -A PSW $(comment "$remarks") -p udp ${_ipt_source} $(REDIRECT $udp_port TPROXY) 2>/dev/null
}
else
msg2="${msg}不代理 UDP"
fi
echolog " - ${msg2}"
}
[ "$_ipv4" != "1" ] && $ip6t_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp -j RETURN 2>/dev/null
$ipt_m -A PSW $(comment "$remarks") ${_ipt_source} -p udp -j RETURN
unset ipt_tmp ipt_j _ipt_source msg msg2 _ipv4
done
unset enabled sid remarks sources use_global_config use_direct_list use_proxy_list use_block_list use_gfw_list chn_list tcp_proxy_mode udp_proxy_mode dns_redirect_port tcp_no_redir_ports udp_no_redir_ports tcp_proxy_drop_ports udp_proxy_drop_ports tcp_redir_ports udp_redir_ports tcp_node udp_node interface
unset tcp_port udp_port tcp_node_remark udp_node_remark _acl_list use_shunt_tcp use_shunt_udp dns_redirect
done
}
[ "$ENABLED_DEFAULT_ACL" == 1 ] && [ "$CLIENT_PROXY" == 1 ] && {
msg="【默认】,"
local ipt_tmp=$ipt_n
[ -n "${is_tproxy}" ] && ipt_tmp=$ipt_m
[ "$TCP_NO_REDIR_PORTS" != "disable" ] && {
add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN"
add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN"
if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]"
else
unset TCP_PROXY_MODE
echolog " - ${msg}不代理所有 TCP 端口"
fi
}
[ "$UDP_NO_REDIR_PORTS" != "disable" ] && {
add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_NO_REDIR_PORTS "-j RETURN"
add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_NO_REDIR_PORTS "-j RETURN"
if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]"
else
unset UDP_PROXY_MODE
echolog " - ${msg}不代理所有 UDP 端口"
fi
}
local DNS_REDIRECT
[ $(config_t_get global dns_redirect "1") = "1" ] && DNS_REDIRECT=53
if ([ -n "$TCP_NODE" ] && [ -n "${TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${UDP_PROXY_MODE}" ]); then
[ -n "${DNS_REDIRECT_PORT}" ] && DNS_REDIRECT=${DNS_REDIRECT_PORT}
else
[ -n "${DIRECT_DNSMASQ_PORT}" ] && DNS_REDIRECT=${DIRECT_DNSMASQ_PORT}
fi
if [ -n "${DNS_REDIRECT}" ]; then
$ipt_m -A PSW $(comment "默认") -p udp --dport 53 -j RETURN
$ip6t_m -A PSW $(comment "默认") -p udp --dport 53 -j RETURN 2>/dev/null
$ipt_m -A PSW $(comment "默认") -p tcp --dport 53 -j RETURN
$ip6t_m -A PSW $(comment "默认") -p tcp --dport 53 -j RETURN 2>/dev/null
$ipt_n -A PSW_DNS $(comment "默认") -p udp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT}
$ip6t_n -A PSW_DNS $(comment "默认") -p udp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT} 2>/dev/null
$ipt_n -A PSW_DNS $(comment "默认") -p tcp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT}
$ip6t_n -A PSW_DNS $(comment "默认") -p tcp --dport 53 -j REDIRECT --to-ports ${DNS_REDIRECT} 2>/dev/null
fi
[ -n "${TCP_PROXY_MODE}" -o -n "${UDP_PROXY_MODE}" ] && {
[ "${USE_BLOCK_LIST}" = "1" ] && $ipt_m -A PSW $(comment "默认") $(dst $IPSET_BLOCK) -j DROP
[ "${USE_DIRECT_LIST}" = "1" ] && $ipt_tmp -A PSW $(comment "默认") $(dst $IPSET_WHITE) -j RETURN
[ "$PROXY_IPV6" == "1" ] && {
[ "${USE_BLOCK_LIST}" = "1" ] && $ip6t_m -A PSW $(comment "默认") $(dst $IPSET_BLOCK6) -j DROP 2>/dev/null
[ "${USE_DIRECT_LIST}" = "1" ] && $ip6t_m -A PSW $(comment "默认") $(dst $IPSET_WHITE6) -j RETURN 2>/dev/null
}
[ "$TCP_PROXY_DROP_PORTS" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && {
add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "-d $FAKE_IP_6 -j DROP"
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK6) -j DROP"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_GFW6) -j DROP"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j DROP")"
[ "${USE_SHUNT_TCP}" = "1" ] && $ip6t_m -A PSW $(comment "默认") -p tcp $TCP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT6) -j DROP"
[ "${TCP_PROXY_MODE}" != "disable" ] && $ip6t_m -A PSW $(comment "默认") -p tcp $TCP_PROXY_DROP_PORTS "-j DROP"
}
add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP"
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")"
[ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS $(dst $IPSET_SHUNT) "-j DROP"
[ "${TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p tcp" $TCP_PROXY_DROP_PORTS "-j DROP"
echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]"
}
[ "$UDP_PROXY_DROP_PORTS" != "disable" ] && {
[ "$PROXY_IPV6" == "1" ] && {
add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-d $FAKE_IP_6 -j DROP"
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK6) -j DROP"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_GFW6) -j DROP"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j DROP")"
[ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT6) -j DROP"
[ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-j DROP"
}
add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP"
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")"
[ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT) -j DROP"
[ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_PROXY_DROP_PORTS "-j DROP"
echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]"
}
}
# 加载TCP默认代理模式
if [ -n "${TCP_PROXY_MODE}" ]; then
[ -n "$TCP_NODE" ] && {
msg2="${msg}使用 TCP 节点[$(config_n_get $TCP_NODE remarks)]"
if [ -n "${is_tproxy}" ]; then
msg2="${msg2}(TPROXY:${TCP_REDIR_PORT})"
ipt_j="-j PSW_RULE"
else
msg2="${msg2}(REDIRECT:${TCP_REDIR_PORT})"
ipt_j="$(REDIRECT $TCP_REDIR_PORT)"
fi
[ "$accept_icmp" = "1" ] && {
$ipt_n -A PSW $(comment "默认") -p icmp -d $FAKE_IP $(REDIRECT)
[ "${USE_PROXY_LIST}" = "1" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_BLACK) $(REDIRECT)
[ "${USE_GFW_LIST}" = "1" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_GFW) $(REDIRECT)
[ "${CHN_LIST}" != "0" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST})
[ "${USE_SHUNT_TCP}" = "1" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(dst $IPSET_SHUNT) $(REDIRECT)
[ "${TCP_PROXY_MODE}" != "disable" ] && $ipt_n -A PSW $(comment "默认") -p icmp $(REDIRECT)
}
[ "$accept_icmpv6" = "1" ] && [ "$PROXY_IPV6" == "1" ] && {
$ip6t_n -A PSW $(comment "默认") -p ipv6-icmp -d $FAKE_IP_6 $(REDIRECT)
[ "${USE_PROXY_LIST}" = "1" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_BLACK6) $(REDIRECT)
[ "${USE_GFW_LIST}" = "1" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_GFW6) $(REDIRECT)
[ "${CHN_LIST}" != "0" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST})
[ "${USE_SHUNT_TCP}" = "1" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(dst $IPSET_SHUNT6) $(REDIRECT)
[ "${TCP_PROXY_MODE}" != "disable" ] && $ip6t_n -A PSW $(comment "默认") -p ipv6-icmp $(REDIRECT)
}
$ipt_tmp -A PSW $(comment "默认") -p tcp -d $FAKE_IP ${ipt_j}
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK) ${ipt_j}"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW) ${ipt_j}"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "${ipt_j}")"
[ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT) ${ipt_j}"
[ "${TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_tmp -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "${ipt_j}"
[ -n "${is_tproxy}" ]&& $ipt_tmp -A PSW $(comment "默认") -p tcp $(REDIRECT $TCP_REDIR_PORT TPROXY)
[ "$PROXY_IPV6" == "1" ] && {
$ip6t_m -A PSW $(comment "默认") -p tcp -d $FAKE_IP_6 -j PSW_RULE
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")"
[ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE"
[ "${TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p tcp" $TCP_REDIR_PORTS "-j PSW_RULE"
$ip6t_m -A PSW $(comment "默认") -p tcp $(REDIRECT $TCP_REDIR_PORT TPROXY)
}
echolog " - ${msg2}"
}
fi
$ipt_n -A PSW $(comment "默认") -p tcp -j RETURN
$ipt_m -A PSW $(comment "默认") -p tcp -j RETURN
$ip6t_m -A PSW $(comment "默认") -p tcp -j RETURN
# 加载UDP默认代理模式
if [ -n "${UDP_PROXY_MODE}" ]; then
[ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ] && {
msg2="${msg}使用 UDP 节点[$(config_n_get $UDP_NODE remarks)](TPROXY:${UDP_REDIR_PORT})"
$ipt_m -A PSW $(comment "默认") -p udp -d $FAKE_IP -j PSW_RULE
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK) -j PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW) -j PSW_RULE"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")"
[ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT) -j PSW_RULE"
[ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "-j PSW_RULE"
$ipt_m -A PSW $(comment "默认") -p udp $(REDIRECT $UDP_REDIR_PORT TPROXY)
[ "$PROXY_IPV6" == "1" ] && {
$ip6t_m -A PSW $(comment "默认") -p udp -d $FAKE_IP_6 -j PSW_RULE
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")"
[ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE"
[ "${UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW $(comment "默认") -p udp" $UDP_REDIR_PORTS "-j PSW_RULE"
$ip6t_m -A PSW $(comment "默认") -p udp $(REDIRECT $UDP_REDIR_PORT TPROXY)
}
echolog " - ${msg2}"
}
fi
$ipt_m -A PSW $(comment "默认") -p udp -j RETURN
$ip6t_m -A PSW $(comment "默认") -p udp -j RETURN
}
}
filter_haproxy() {
for item in ${haproxy_items}; do
local ip=$(get_host_ip ipv4 $(echo $item | awk -F ":" '{print $1}') 1)
ipset -q add $IPSET_VPS $ip
done
echolog " - [$?]加入负载均衡的节点到ipset[$IPSET_VPS]直连完成"
}
filter_vpsip() {
uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([0-9]{1,3}[\.]){3}[0-9]{1,3}" | grep -v "^127\.0\.0\.1$" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_VPS &/g" | awk '{print $0} END{print "COMMIT"}' | ipset -! -R
echolog " - [$?]加入所有IPv4节点到ipset[$IPSET_VPS]直连完成"
uci show $CONFIG | grep -E "(.address=|.download_address=)" | cut -d "'" -f 2 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_VPS6 &/g" | awk '{print $0} END{print "COMMIT"}' | ipset -! -R
echolog " - [$?]加入所有IPv6节点到ipset[$IPSET_VPS6]直连完成"
}
filter_server_port() {
local address=${1}
local port=${2}
local stream=${3}
stream=$(echo ${3} | tr 'A-Z' 'a-z')
local _is_tproxy ipt_tmp
ipt_tmp=$ipt_n
_is_tproxy=${is_tproxy}
[ "$stream" == "udp" ] && _is_tproxy="TPROXY"
[ -n "${_is_tproxy}" ] && ipt_tmp=$ipt_m
for _ipt in 4 6; do
[ "$_ipt" == "4" ] && _ipt=$ipt_tmp
[ "$_ipt" == "6" ] && _ipt=$ip6t_m
$_ipt -n -L PSW_OUTPUT | grep -q "${address}:${port}"
if [ $? -ne 0 ]; then
$_ipt -I PSW_OUTPUT $(comment "${address}:${port}") -p $stream -d $address --dport $port -j RETURN 2>/dev/null
fi
done
}
filter_node() {
local node=${1}
local stream=${2}
if [ -n "$node" ]; then
local address=$(config_n_get $node address)
local port=$(config_n_get $node port)
[ -z "$address" ] && [ -z "$port" ] && {
return 1
}
filter_server_port $address $port $stream
filter_server_port $address $port $stream
fi
}
filter_direct_node_list() {
[ ! -s "$TMP_PATH/direct_node_list" ] && return
for _node_id in $(cat $TMP_PATH/direct_node_list | awk '!seen[$0]++'); do
filter_node "$_node_id" TCP
filter_node "$_node_id" UDP
unset _node_id
done
}
add_firewall_rule() {
echolog "开始加载防火墙规则..."
ipset -! create $IPSET_LOCAL nethash maxelem 1048576
ipset -! create $IPSET_LAN nethash maxelem 1048576
ipset -! create $IPSET_VPS nethash maxelem 1048576
ipset -! create $IPSET_SHUNT nethash maxelem 1048576 timeout 172800
ipset -! create $IPSET_GFW nethash maxelem 1048576 timeout 172800
ipset -! create $IPSET_CHN nethash maxelem 1048576 timeout 172800
ipset -! create $IPSET_BLACK nethash maxelem 1048576 timeout 172800
ipset -! create $IPSET_WHITE nethash maxelem 1048576 timeout 172800
ipset -! create $IPSET_BLOCK nethash maxelem 1048576 timeout 172800
ipset -! create $IPSET_LOCAL6 nethash family inet6 maxelem 1048576
ipset -! create $IPSET_LAN6 nethash family inet6 maxelem 1048576
ipset -! create $IPSET_VPS6 nethash family inet6 maxelem 1048576
ipset -! create $IPSET_SHUNT6 nethash family inet6 maxelem 1048576 timeout 172800
ipset -! create $IPSET_GFW6 nethash family inet6 maxelem 1048576 timeout 172800
ipset -! create $IPSET_CHN6 nethash family inet6 maxelem 1048576 timeout 172800
ipset -! create $IPSET_BLACK6 nethash family inet6 maxelem 1048576 timeout 172800
ipset -! create $IPSET_WHITE6 nethash family inet6 maxelem 1048576 timeout 172800
ipset -! create $IPSET_BLOCK6 nethash family inet6 maxelem 1048576 timeout 172800
cat $RULES_PATH/chnroute | tr -s '\n' | grep -v "^#" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_CHN &/g" -e "s/$/ timeout 0/g" | ipset -! -R
cat $RULES_PATH/chnroute6 | tr -s '\n' | grep -v "^#" | sed -e "/^$/d" | sed -e "s/^/add $IPSET_CHN6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
#导入规则列表、分流规则中的IP列表
local USE_SHUNT_NODE=0
local USE_PROXY_LIST_ALL=${USE_PROXY_LIST}
local USE_DIRECT_LIST_ALL=${USE_DIRECT_LIST}
local USE_BLOCK_LIST_ALL=${USE_BLOCK_LIST}
local _TCP_NODE=$(config_t_get global tcp_node)
local _UDP_NODE=$(config_t_get global udp_node)
local USE_GEOVIEW=$(config_t_get global_rules enable_geoview)
[ -n "$_TCP_NODE" ] && [ "$(config_n_get $_TCP_NODE protocol)" = "_shunt" ] && USE_SHUNT_TCP=1 && USE_SHUNT_NODE=1
[ -n "$_UDP_NODE" ] && [ "$(config_n_get $_UDP_NODE protocol)" = "_shunt" ] && USE_SHUNT_UDP=1 && USE_SHUNT_NODE=1
[ "$_UDP_NODE" = "tcp" ] && USE_SHUNT_UDP=$USE_SHUNT_TCP
for acl_section in $(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do
[ "$(config_n_get $acl_section enabled)" != "1" ] && continue
[ "$(config_n_get $acl_section use_global_config 0)" != "1" ] && {
[ "$(config_n_get $acl_section use_direct_list 1)" = "1" ] && USE_PROXY_LIST_ALL=1
[ "$(config_n_get $acl_section use_proxy_list 1)" = "1" ] && USE_DIRECT_LIST_ALL=1
[ "$(config_n_get $acl_section use_block_list 1)" = "1" ] && USE_BLOCK_LIST_ALL=1
}
for _node in $(config_n_get $acl_section tcp_node) $(config_n_get $acl_section udp_node); do
local node_protocol=$(config_n_get $_node protocol)
[ "$node_protocol" = "_shunt" ] && { USE_SHUNT_NODE=1; break; }
done
done
#直连列表
[ "$USE_DIRECT_LIST_ALL" = "1" ] && {
cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_WHITE &/g" -e "s/$/ timeout 0/g" | ipset -! -R
cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_WHITE6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
[ "$USE_GEOVIEW" = "1" ] && {
local GEOIP_CODE=$(cat $RULES_PATH/direct_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_WHITE &/g" -e "s/$/ timeout 0/g" | ipset -! -R
get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_WHITE6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
echolog " - [$?]解析并加入[直连列表] GeoIP 到 IPSET 完成"
fi
}
}
#代理列表
[ "$USE_PROXY_LIST_ALL" = "1" ] && {
cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLACK &/g" -e "s/$/ timeout 0/g" | ipset -! -R
cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLACK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
[ "$USE_GEOVIEW" = "1" ] && {
local GEOIP_CODE=$(cat $RULES_PATH/proxy_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLACK &/g" -e "s/$/ timeout 0/g" | ipset -! -R
get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLACK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
echolog " - [$?]解析并加入[代理列表] GeoIP 到 IPSET 完成"
fi
}
}
#屏蔽列表
[ "$USE_BLOCK_LIST_ALL" = "1" ] && {
cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLOCK &/g" -e "s/$/ timeout 0/g" | ipset -! -R
cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLOCK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
[ "$USE_GEOVIEW" = "1" ] && {
local GEOIP_CODE=$(cat $RULES_PATH/block_ip | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_BLOCK &/g" -e "s/$/ timeout 0/g" | ipset -! -R
get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_BLOCK6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
echolog " - [$?]解析并加入[屏蔽列表] GeoIP 到 IPSET 完成"
fi
}
}
#分流列表
[ "$USE_SHUNT_NODE" = "1" ] && {
local GEOIP_CODE=""
local shunt_ids=$(uci show $CONFIG | grep "=shunt_rules" | awk -F '.' '{print $2}' | awk -F '=' '{print $1}')
for shunt_id in $shunt_ids; do
config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_SHUNT &/g" -e "s/$/ timeout 0/g" | ipset -! -R
config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | grep -v "^#" | sed -e "/^$/d" | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_SHUNT6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
[ "$USE_GEOVIEW" = "1" ] && {
local geoip_code=$(config_n_get $shunt_id ip_list | tr -s "\r\n" "\n" | sed -e "/^$/d" | grep -E "^geoip:" | grep -v "^geoip:private" | sed -E 's/^geoip:(.*)/\1/' | sed ':a;N;$!ba;s/\n/,/g')
[ -n "$geoip_code" ] && GEOIP_CODE="${GEOIP_CODE:+$GEOIP_CODE,}$geoip_code"
}
done
if [ -n "$GEOIP_CODE" ] && type geoview &> /dev/null; then
get_geoip $GEOIP_CODE ipv4 | grep -E "(\.((2(5[0-5]|[0-4][0-9]))|[0-1]?[0-9]{1,2})){3}" | sed -e "s/^/add $IPSET_SHUNT &/g" -e "s/$/ timeout 0/g" | ipset -! -R
get_geoip $GEOIP_CODE ipv6 | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | sed -e "s/^/add $IPSET_SHUNT6 &/g" -e "s/$/ timeout 0/g" | ipset -! -R
echolog " - [$?]解析并加入[分流节点] GeoIP 到 IPSET 完成"
fi
}
ipset -! -R <<-EOF
$(ip address show | grep -w "inet" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/^/add $IPSET_LOCAL /")
EOF
ipset -! -R <<-EOF
$(ip address show | grep -w "inet6" | awk '{print $2}' | awk -F '/' '{print $1}' | sed -e "s/^/add $IPSET_LOCAL6 /")
EOF
#局域网IP列表
ipset -! -R <<-EOF
$(gen_lanlist | sed -e "s/^/add $IPSET_LAN /")
EOF
ipset -! -R <<-EOF
$(gen_lanlist_6 | sed -e "s/^/add $IPSET_LAN6 /")
EOF
# 忽略特殊IP段
local lan_ifname lan_ip
lan_ifname=$(uci -q -p /tmp/state get network.lan.ifname)
[ -n "$lan_ifname" ] && {
lan_ip=$(ip address show $lan_ifname | grep -w "inet" | awk '{print $2}')
lan_ip6=$(ip address show $lan_ifname | grep -w "inet6" | awk '{print $2}')
#echolog "本机IPv4网段互访直连:${lan_ip}"
#echolog "本机IPv6网段互访直连:${lan_ip6}"
[ -n "$lan_ip" ] && ipset -! -R <<-EOF
$(echo $lan_ip | sed -e "s/ /\n/g" | sed -e "s/^/add $IPSET_LAN /")
EOF
[ -n "$lan_ip6" ] && ipset -! -R <<-EOF
$(echo $lan_ip6 | sed -e "s/ /\n/g" | sed -e "s/^/add $IPSET_LAN6 /")
EOF
}
[ -n "$ISP_DNS" ] && {
#echolog "处理 ISP DNS 例外..."
for ispip in $ISP_DNS; do
ipset -! add $IPSET_WHITE $ispip timeout 0
echolog " - [$?]追加ISP IPv4 DNS到白名单:${ispip}"
done
}
[ -n "$ISP_DNS6" ] && {
#echolog "处理 ISP IPv6 DNS 例外..."
for ispip6 in $ISP_DNS6; do
ipset -! add $IPSET_WHITE6 $ispip6 timeout 0
echolog " - [$?]追加ISP IPv6 DNS到白名单:${ispip6}"
done
}
# 过滤所有节点IP
filter_vpsip > /dev/null 2>&1 &
# filter_haproxy > /dev/null 2>&1 &
accept_icmp=$(config_t_get global_forwarding accept_icmp 0)
accept_icmpv6=$(config_t_get global_forwarding accept_icmpv6 0)
if [ "${TCP_PROXY_WAY}" = "redirect" ]; then
unset is_tproxy
elif [ "${TCP_PROXY_WAY}" = "tproxy" ]; then
is_tproxy="TPROXY"
fi
$ipt_n -N PSW
$ipt_n -A PSW $(dst $IPSET_LAN) -j RETURN
$ipt_n -A PSW $(dst $IPSET_VPS) -j RETURN
WAN_IP=$(get_wan_ip)
[ ! -z "${WAN_IP}" ] && $ipt_n -A PSW $(comment "WAN_IP_RETURN") -d "${WAN_IP}" -j RETURN
[ "$accept_icmp" = "1" ] && insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p icmp -j PSW"
[ -z "${is_tproxy}" ] && insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p tcp -j PSW"
$ipt_n -N PSW_OUTPUT
$ipt_n -A PSW_OUTPUT $(dst $IPSET_LAN) -j RETURN
$ipt_n -A PSW_OUTPUT $(dst $IPSET_VPS) -j RETURN
[ "${USE_DIRECT_LIST}" = "1" ] && $ipt_n -A PSW_OUTPUT $(dst $IPSET_WHITE) -j RETURN
$ipt_n -A PSW_OUTPUT -m mark --mark 0xff -j RETURN
$ipt_n -N PSW_DNS
if [ $(config_t_get global dns_redirect "1") = "0" ]; then
#Only hijack when dest address is local IP
$ipt_n -I PREROUTING $(dst $IPSET_LOCAL) -j PSW_DNS
else
$ipt_n -I PREROUTING 1 -j PSW_DNS
fi
$ipt_m -N PSW_DIVERT
$ipt_m -A PSW_DIVERT -j MARK --set-mark 1
$ipt_m -A PSW_DIVERT -j ACCEPT
$ipt_m -N PSW_RULE
$ipt_m -A PSW_RULE -j CONNMARK --restore-mark
$ipt_m -A PSW_RULE -m mark --mark 1 -j RETURN
$ipt_m -A PSW_RULE -p tcp -m tcp --tcp-flags FIN,SYN,RST,ACK SYN -j MARK --set-xmark 1
$ipt_m -A PSW_RULE -p udp -m conntrack --ctstate NEW -j MARK --set-xmark 1
$ipt_m -A PSW_RULE -j CONNMARK --save-mark
$ipt_m -N PSW
$ipt_m -A PSW $(dst $IPSET_LAN) -j RETURN
$ipt_m -A PSW $(dst $IPSET_VPS) -j RETURN
[ ! -z "${WAN_IP}" ] && {
$ipt_m -A PSW $(comment "WAN_IP_RETURN") -d "${WAN_IP}" -j RETURN
echolog " - [$?]追加WAN IP到iptables:${WAN_IP}"
}
unset WAN_IP
insert_rule_before "$ipt_m" "PREROUTING" "mwan3" "-j PSW"
insert_rule_before "$ipt_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT"
$ipt_m -N PSW_OUTPUT
$ipt_m -A PSW_OUTPUT $(dst $IPSET_LAN) -j RETURN
$ipt_m -A PSW_OUTPUT $(dst $IPSET_VPS) -j RETURN
[ -n "$IPT_APPEND_DNS" ] && {
local local_dns dns_address dns_port
for local_dns in $(echo $IPT_APPEND_DNS | tr ',' ' '); do
dns_address=$(echo "$local_dns" | sed -E 's/(@|\[)?([0-9a-fA-F:.]+)(@|#|$).*/\2/')
dns_port=$(echo "$local_dns" | sed -nE 's/.*#([0-9]+)$/\1/p')
if echo "$dns_address" | grep -q -v ':'; then
$ipt_m -A PSW_OUTPUT -p udp -d ${dns_address} --dport ${dns_port:-53} -j RETURN
$ipt_m -A PSW_OUTPUT -p tcp -d ${dns_address} --dport ${dns_port:-53} -j RETURN
echolog " - [$?]追加直连DNS到iptables:${dns_address}:${dns_port:-53}"
else
$ip6t_m -A PSW_OUTPUT -p udp -d ${dns_address} --dport ${dns_port:-53} -j RETURN
$ip6t_m -A PSW_OUTPUT -p tcp -d ${dns_address} --dport ${dns_port:-53} -j RETURN
echolog " - [$?]追加直连DNS到iptables:[${dns_address}]:${dns_port:-53}"
fi
done
}
[ "${USE_BLOCK_LIST}" = "1" ] && $ipt_m -A PSW_OUTPUT $(dst $IPSET_BLOCK) -j DROP
[ "${USE_DIRECT_LIST}" = "1" ] && $ipt_m -A PSW_OUTPUT $(dst $IPSET_WHITE) -j RETURN
$ipt_m -A PSW_OUTPUT -m mark --mark 0xff -j RETURN
ip rule add fwmark 1 lookup 100
ip route add local 0.0.0.0/0 dev lo table 100
[ "$accept_icmpv6" = "1" ] && {
$ip6t_n -N PSW
$ip6t_n -A PSW $(dst $IPSET_LAN6) -j RETURN
$ip6t_n -A PSW $(dst $IPSET_VPS6) -j RETURN
$ip6t_n -A PREROUTING -p ipv6-icmp -j PSW
$ip6t_n -N PSW_OUTPUT
$ip6t_n -A PSW_OUTPUT $(dst $IPSET_LAN6) -j RETURN
$ip6t_n -A PSW_OUTPUT $(dst $IPSET_VPS6) -j RETURN
[ "${USE_DIRECT_LIST}" = "1" ] && $ip6t_n -A PSW_OUTPUT $(dst $IPSET_WHITE6) -j RETURN
$ip6t_n -A PSW_OUTPUT -m mark --mark 0xff -j RETURN
}
$ip6t_n -N PSW_DNS
if [ $(config_t_get global dns_redirect "1") = "0" ]; then
#Only hijack when dest address is local IP
$ip6t_n -I PREROUTING $(dst $IPSET_LOCAL6) -j PSW_DNS
else
$ip6t_n -I PREROUTING 1 -j PSW_DNS
fi
$ip6t_m -N PSW_DIVERT
$ip6t_m -A PSW_DIVERT -j MARK --set-mark 1
$ip6t_m -A PSW_DIVERT -j ACCEPT
$ip6t_m -N PSW_RULE
$ip6t_m -A PSW_RULE -j CONNMARK --restore-mark
$ip6t_m -A PSW_RULE -m mark --mark 1 -j RETURN
$ip6t_m -A PSW_RULE -p tcp -m tcp --tcp-flags FIN,SYN,RST,ACK SYN -j MARK --set-xmark 1
$ip6t_m -A PSW_RULE -p udp -m conntrack --ctstate NEW -j MARK --set-xmark 1
$ip6t_m -A PSW_RULE -j CONNMARK --save-mark
$ip6t_m -N PSW
$ip6t_m -A PSW $(dst $IPSET_LAN6) -j RETURN
$ip6t_m -A PSW $(dst $IPSET_VPS6) -j RETURN
WAN6_IP=$(get_wan6_ip)
[ ! -z "${WAN6_IP}" ] && $ip6t_m -A PSW $(comment "WAN6_IP_RETURN") -d ${WAN6_IP} -j RETURN
unset WAN6_IP
insert_rule_before "$ip6t_m" "PREROUTING" "mwan3" "-j PSW"
insert_rule_before "$ip6t_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT"
$ip6t_m -N PSW_OUTPUT
$ip6t_m -A PSW_OUTPUT -m mark --mark 0xff -j RETURN
$ip6t_m -A PSW_OUTPUT $(dst $IPSET_LAN6) -j RETURN
$ip6t_m -A PSW_OUTPUT $(dst $IPSET_VPS6) -j RETURN
[ "${USE_BLOCK_LIST}" = "1" ] && $ip6t_m -A PSW_OUTPUT $(dst $IPSET_BLOCK6) -j DROP
[ "${USE_DIRECT_LIST}" = "1" ] && $ip6t_m -A PSW_OUTPUT $(dst $IPSET_WHITE6) -j RETURN
ip -6 rule add fwmark 1 table 100
ip -6 route add local ::/0 dev lo table 100
[ "$TCP_UDP" = "1" ] && [ -z "$UDP_NODE" ] && UDP_NODE=$TCP_NODE
[ "$ENABLED_DEFAULT_ACL" == 1 ] && {
local ipt_tmp=$ipt_n
if [ -n "${is_tproxy}" ]; then
ipt_tmp=$ipt_m
ipt_j="-j PSW_RULE"
else
ipt_j="$(REDIRECT $TCP_REDIR_PORT)"
fi
msg="【路由器本机】,"
[ "$TCP_NO_REDIR_PORTS" != "disable" ] && {
add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN"
add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_NO_REDIR_PORTS "-j RETURN"
if ! has_1_65535 "$TCP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 TCP 端口[${TCP_NO_REDIR_PORTS}]"
else
unset LOCALHOST_TCP_PROXY_MODE
echolog " - ${msg}不代理所有 TCP 端口"
fi
}
[ "$UDP_NO_REDIR_PORTS" != "disable" ] && {
add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_NO_REDIR_PORTS "-j RETURN"
add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_NO_REDIR_PORTS "-j RETURN"
if ! has_1_65535 "$UDP_NO_REDIR_PORTS"; then
echolog " - ${msg}不代理 UDP 端口[${UDP_NO_REDIR_PORTS}]"
else
unset LOCALHOST_UDP_PROXY_MODE
echolog " - ${msg}不代理所有 UDP 端口"
fi
}
if ([ -n "$TCP_NODE" ] && [ -n "${LOCALHOST_TCP_PROXY_MODE}" ]) || ([ -n "$UDP_NODE" ] && [ -n "${LOCALHOST_UDP_PROXY_MODE}" ]); then
[ -n "$DNS_REDIRECT_PORT" ] && {
$ipt_n -A OUTPUT $(comment "PSW_DNS") -p udp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT
$ip6t_n -A OUTPUT $(comment "PSW_DNS") -p udp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT 2>/dev/null
$ipt_n -A OUTPUT $(comment "PSW_DNS") -p tcp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT
$ip6t_n -A OUTPUT $(comment "PSW_DNS") -p tcp -o lo --dport 53 -j REDIRECT --to-ports $DNS_REDIRECT_PORT 2>/dev/null
}
fi
[ -n "${LOCALHOST_TCP_PROXY_MODE}" -o -n "${LOCALHOST_UDP_PROXY_MODE}" ] && {
[ "$TCP_PROXY_DROP_PORTS" != "disable" ] && {
add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP"
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")"
[ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT) -j DROP"
[ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p tcp" $TCP_PROXY_DROP_PORTS "-j DROP"
echolog " - ${msg}屏蔽代理 TCP 端口[${TCP_PROXY_DROP_PORTS}]"
}
[ "$UDP_PROXY_DROP_PORTS" != "disable" ] && {
add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "-d $FAKE_IP -j DROP"
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_BLACK) -j DROP"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_GFW) -j DROP"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j DROP")"
[ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "$(dst $IPSET_SHUNT) -j DROP"
[ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_PROXY_DROP_PORTS "-j DROP"
echolog " - ${msg}屏蔽代理 UDP 端口[${UDP_PROXY_DROP_PORTS}]"
}
}
# 加载路由器自身代理 TCP
if [ -n "$TCP_NODE" ]; then
_proxy_tcp_access() {
[ -n "${2}" ] || return 0
if echo "${2}" | grep -q -v ':'; then
ipset -q test $IPSET_LAN ${2}
[ $? -eq 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问"
return 0
}
if [ -z "${is_tproxy}" ]; then
$ipt_n -I PSW_OUTPUT -p tcp -d ${2} --dport ${3} $(REDIRECT $TCP_REDIR_PORT)
else
$ipt_m -I PSW_OUTPUT -p tcp -d ${2} --dport ${3} -j PSW_RULE
$ipt_m -I PSW $(comment "本机") -p tcp -i lo -d ${2} --dport ${3} $(REDIRECT $TCP_REDIR_PORT TPROXY)
fi
echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 TCP 转发链"
else
ipset -q test $IPSET_LAN6 ${2}
[ $? -eq 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 TCP 代理转发对该服务器 TCP/${3} 端口的访问"
return 0
}
$ip6t_m -I PSW_OUTPUT -p tcp -d ${2} --dport ${3} -j PSW_RULE
$ip6t_m -I PSW $(comment "本机") -p tcp -i lo -d ${2} --dport ${3} $(REDIRECT $TCP_REDIR_PORT TPROXY)
echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 TCP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!"
fi
}
[ -n "${TCP_PROXY_DNS}" ] && hosts_foreach REMOTE_DNS _proxy_tcp_access 53
[ "$accept_icmp" = "1" ] && {
$ipt_n -A OUTPUT -p icmp -j PSW_OUTPUT
$ipt_n -A PSW_OUTPUT -p icmp -d $FAKE_IP $(REDIRECT)
[ "${USE_PROXY_LIST}" = "1" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_BLACK) $(REDIRECT)
[ "${USE_GFW_LIST}" = "1" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_GFW) $(REDIRECT)
[ "${CHN_LIST}" != "0" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST})
[ "${USE_SHUNT_TCP}" = "1" ] && $ipt_n -A PSW_OUTPUT -p icmp $(dst $IPSET_SHUNT) $(REDIRECT)
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && $ipt_n -A PSW_OUTPUT -p icmp $(REDIRECT)
}
[ "$accept_icmpv6" = "1" ] && {
$ip6t_n -A OUTPUT -p ipv6-icmp -j PSW_OUTPUT
$ip6t_n -A PSW_OUTPUT -p ipv6-icmp -d $FAKE_IP_6 $(REDIRECT)
[ "${USE_PROXY_LIST}" = "1" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_BLACK6) $(REDIRECT)
[ "${USE_GFW_LIST}" = "1" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_GFW6) $(REDIRECT)
[ "${CHN_LIST}" != "0" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST})
[ "${USE_SHUNT_TCP}" = "1" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(dst $IPSET_SHUNT6) $(REDIRECT)
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && [ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && $ip6t_n -A PSW_OUTPUT -p ipv6-icmp $(REDIRECT)
}
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && {
$ipt_tmp -A PSW_OUTPUT -p tcp -d $FAKE_IP ${ipt_j}
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK) ${ipt_j}"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW) ${ipt_j}"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "${ipt_j}")"
[ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT) ${ipt_j}"
[ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_tmp -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "${ipt_j}"
[ -n "${is_tproxy}" ] && $ipt_m -A PSW $(comment "本机") -p tcp -i lo $(REDIRECT $TCP_REDIR_PORT TPROXY)
}
[ -z "${is_tproxy}" ] && $ipt_n -A OUTPUT -p tcp -j PSW_OUTPUT
[ -n "${is_tproxy}" ] && {
$ipt_m -A PSW $(comment "本机") -p tcp -i lo -j RETURN
insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p tcp -j PSW_OUTPUT"
}
[ "$PROXY_IPV6" == "1" ] && {
[ -n "${LOCALHOST_TCP_PROXY_MODE}" ] && {
$ip6t_m -A PSW_OUTPUT -p tcp -d $FAKE_IP_6 -j PSW_RULE
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")"
[ "${USE_SHUNT_TCP}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE"
[ "${LOCALHOST_TCP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p tcp" $TCP_REDIR_PORTS "-j PSW_RULE"
$ip6t_m -A PSW $(comment "本机") -p tcp -i lo $(REDIRECT $TCP_REDIR_PORT TPROXY)
}
$ip6t_m -A PSW $(comment "本机") -p tcp -i lo -j RETURN
insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p tcp -j PSW_OUTPUT"
}
fi
# 加载路由器自身代理 UDP
if [ -n "$UDP_NODE" -o "$TCP_UDP" = "1" ]; then
_proxy_udp_access() {
[ -n "${2}" ] || return 0
if echo "${2}" | grep -q -v ':'; then
ipset -q test $IPSET_LAN ${2}
[ $? == 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问"
return 0
}
$ipt_m -I PSW_OUTPUT -p udp -d ${2} --dport ${3} -j PSW_RULE
$ipt_m -I PSW $(comment "本机") -p udp -i lo -d ${2} --dport ${3} $(REDIRECT $UDP_REDIR_PORT TPROXY)
echolog " - [$?]将上游 DNS 服务器 ${2}:${3} 加入到路由器自身代理的 UDP 转发链"
else
ipset -q test $IPSET_LAN6 ${2}
[ $? == 0 ] && {
echolog " - 上游 DNS 服务器 ${2} 已在直接访问的列表中,不强制向 UDP 代理转发对该服务器 UDP/${3} 端口的访问"
return 0
}
$ip6t_m -I PSW_OUTPUT -p udp -d ${2} --dport ${3} -j PSW_RULE
$ip6t_m -I PSW $(comment "本机") -p udp -i lo -d ${2} --dport ${3} $(REDIRECT $UDP_REDIR_PORT TPROXY)
echolog " - [$?]将上游 DNS 服务器 [${2}]:${3} 加入到路由器自身代理的 UDP 转发链,请确保您的节点支持IPv6,并开启IPv6透明代理!"
fi
}
[ -n "${UDP_PROXY_DNS}" ] && hosts_foreach REMOTE_DNS _proxy_udp_access 53
[ -n "${LOCALHOST_UDP_PROXY_MODE}" ] && {
$ipt_m -A PSW_OUTPUT -p udp -d $FAKE_IP -j PSW_RULE
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK) -j PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW) -j PSW_RULE"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")"
[ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT) -j PSW_RULE"
[ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ipt_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "-j PSW_RULE"
$ipt_m -A PSW $(comment "本机") -p udp -i lo $(REDIRECT $UDP_REDIR_PORT TPROXY)
}
$ipt_m -A PSW $(comment "本机") -p udp -i lo -j RETURN
insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p udp -j PSW_OUTPUT"
[ "$PROXY_IPV6" == "1" ] && {
[ -n "$LOCALHOST_UDP_PROXY_MODE" ] && {
$ip6t_m -A PSW_OUTPUT -p udp -d $FAKE_IP_6 -j PSW_RULE
[ "${USE_PROXY_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_BLACK6) -j PSW_RULE"
[ "${USE_GFW_LIST}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_GFW6) -j PSW_RULE"
[ "${CHN_LIST}" != "0" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_CHN6) $(get_jump_ipt ${CHN_LIST} "-j PSW_RULE")"
[ "${USE_SHUNT_UDP}" = "1" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "$(dst $IPSET_SHUNT6) -j PSW_RULE"
[ "${LOCALHOST_UDP_PROXY_MODE}" != "disable" ] && add_port_rules "$ip6t_m -A PSW_OUTPUT -p udp" $UDP_REDIR_PORTS "-j PSW_RULE"
$ip6t_m -A PSW $(comment "本机") -p udp -i lo $(REDIRECT $UDP_REDIR_PORT TPROXY)
}
$ip6t_m -A PSW $(comment "本机") -p udp -i lo -j RETURN
insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -p udp -j PSW_OUTPUT"
}
fi
$ipt_m -I OUTPUT $(comment "mangle-OUTPUT-PSW") -o lo -j RETURN
insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -m mark --mark 1 -j RETURN"
$ip6t_m -I OUTPUT $(comment "mangle-OUTPUT-PSW") -o lo -j RETURN
insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "$(comment mangle-OUTPUT-PSW) -m mark --mark 1 -j RETURN"
}
# 加载ACLS
load_acl
[ -d "${TMP_IFACE_PATH}" ] && {
for iface in $(ls ${TMP_IFACE_PATH}); do
$ipt_n -I PSW_OUTPUT -o $iface -j RETURN
$ipt_m -I PSW_OUTPUT -o $iface -j RETURN
done
}
$ipt_n -I PREROUTING $(comment "PSW") -m mark --mark 1 -j RETURN
$ip6t_n -I PREROUTING $(comment "PSW") -m mark --mark 1 -j RETURN
filter_direct_node_list > /dev/null 2>&1 &
echolog "防火墙规则加载完成!"
}
del_firewall_rule() {
for ipt in "$ipt_n" "$ipt_m" "$ip6t_n" "$ip6t_m"; do
for chain in "PREROUTING" "OUTPUT"; do
for i in $(seq 1 $($ipt -nL $chain | grep -c PSW)); do
local index=$($ipt --line-number -nL $chain | grep PSW | head -1 | awk '{print $1}')
$ipt -D $chain $index 2>/dev/null
done
done
for chain in "PSW" "PSW_OUTPUT" "PSW_DIVERT" "PSW_DNS" "PSW_RULE"; do
$ipt -F $chain 2>/dev/null
$ipt -X $chain 2>/dev/null
done
done
ip rule del fwmark 1 lookup 100 2>/dev/null
ip route del local 0.0.0.0/0 dev lo table 100 2>/dev/null
ip -6 rule del fwmark 1 table 100 2>/dev/null
ip -6 route del local ::/0 dev lo table 100 2>/dev/null
destroy_ipset $IPSET_LOCAL
destroy_ipset $IPSET_LAN
destroy_ipset $IPSET_VPS
destroy_ipset $IPSET_SHUNT
#destroy_ipset $IPSET_GFW
#destroy_ipset $IPSET_CHN
#destroy_ipset $IPSET_BLACK
destroy_ipset $IPSET_BLOCK
destroy_ipset $IPSET_WHITE
destroy_ipset $IPSET_LOCAL6
destroy_ipset $IPSET_LAN6
destroy_ipset $IPSET_VPS6
destroy_ipset $IPSET_SHUNT6
#destroy_ipset $IPSET_GFW6
#destroy_ipset $IPSET_CHN6
#destroy_ipset $IPSET_BLACK6
destroy_ipset $IPSET_BLOCK6
destroy_ipset $IPSET_WHITE6
$DIR/app.sh echolog "删除 iptables 规则完成。"
}
flush_ipset() {
$DIR/app.sh echolog "清空 IPSet。"
for _name in $(ipset list | grep "Name: " | grep "passwall_" | awk '{print $2}'); do
destroy_ipset ${_name}
done
}
flush_include() {
echo '#!/bin/sh' >$FWI
}
gen_include() {
flush_include
extract_rules() {
local _ipt="${ipt}"
[ "$1" == "6" ] && _ipt="${ip6t}"
[ -z "${_ipt}" ] && return
echo "*$2"
${_ipt}-save -t $2 | grep "PSW" | grep -v "\-j PSW$" | grep -v "mangle\-OUTPUT\-PSW" | grep -v "socket \-j PSW_DIVERT$" | sed -e "s/^-A \(OUTPUT\|PREROUTING\)/-I \1 1/"
echo 'COMMIT'
}
local __ipt=""
[ -n "${ipt}" ] && {
__ipt=$(cat <<- EOF
mangle_output_psw=\$(${ipt}-save -t mangle | grep "PSW" | grep "mangle\-OUTPUT\-PSW" | sed "s#-A OUTPUT ##g")
$ipt-save -c | grep -v "PSW" | $ipt-restore -c
$ipt-restore -n <<-EOT
$(extract_rules 4 nat)
$(extract_rules 4 mangle)
EOT
echo "\${mangle_output_psw}" | while read line; do
\$(${MY_PATH} insert_rule_before "$ipt_m" "OUTPUT" "mwan3" "\${line}")
done
[ "$accept_icmp" = "1" ] && \$(${MY_PATH} insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p icmp -j PSW")
[ -z "${is_tproxy}" ] && \$(${MY_PATH} insert_rule_after "$ipt_n" "PREROUTING" "prerouting_rule" "-p tcp -j PSW")
\$(${MY_PATH} insert_rule_before "$ipt_m" "PREROUTING" "mwan3" "-j PSW")
\$(${MY_PATH} insert_rule_before "$ipt_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT")
WAN_IP=\$(${MY_PATH} get_wan_ip)
PR_INDEX=\$(${MY_PATH} RULE_LAST_INDEX "$ipt_n" PSW WAN_IP_RETURN -1)
if [ \$PR_INDEX -ge 0 ]; then
[ ! -z "\${WAN_IP}" ] && $ipt_n -R PSW \$PR_INDEX $(comment "WAN_IP_RETURN") -d "\${WAN_IP}" -j RETURN
fi
PR_INDEX=\$(${MY_PATH} RULE_LAST_INDEX "$ipt_m" PSW WAN_IP_RETURN -1)
if [ \$PR_INDEX -ge 0 ]; then
[ ! -z "\${WAN_IP}" ] && $ipt_m -R PSW \$PR_INDEX $(comment "WAN_IP_RETURN") -d "\${WAN_IP}" -j RETURN
fi
EOF
)
}
local __ip6t=""
[ -n "${ip6t}" ] && {
__ip6t=$(cat <<- EOF
mangle_output_psw=\$(${ip6t}-save -t mangle | grep "PSW" | grep "mangle\-OUTPUT\-PSW" | sed "s#-A OUTPUT ##g")
$ip6t-save -c | grep -v "PSW" | $ip6t-restore -c
$ip6t-restore -n <<-EOT
$(extract_rules 6 nat)
$(extract_rules 6 mangle)
EOT
echo "\${mangle_output_psw}" | while read line; do
\$(${MY_PATH} insert_rule_before "$ip6t_m" "OUTPUT" "mwan3" "\${line}")
done
[ "$accept_icmpv6" = "1" ] && $ip6t_n -A PREROUTING -p ipv6-icmp -j PSW
\$(${MY_PATH} insert_rule_before "$ip6t_m" "PREROUTING" "mwan3" "-j PSW")
\$(${MY_PATH} insert_rule_before "$ip6t_m" "PREROUTING" "PSW" "-p tcp -m socket -j PSW_DIVERT")
PR_INDEX=\$(${MY_PATH} RULE_LAST_INDEX "$ip6t_m" PSW WAN6_IP_RETURN -1)
if [ \$PR_INDEX -ge 0 ]; then
WAN6_IP=\$(${MY_PATH} get_wan6_ip)
[ ! -z "\${WAN6_IP}" ] && $ip6t_m -R PSW \$PR_INDEX $(comment "WAN6_IP_RETURN") -d "\${WAN6_IP}" -j RETURN
fi
EOF
)
}
cat <<-EOF >> $FWI
${__ipt}
${__ip6t}
return 0
EOF
return 0
}
get_ipt_bin() {
echo $ipt
}
get_ip6t_bin() {
echo $ip6t
}
start() {
[ "$ENABLED_DEFAULT_ACL" == 0 -a "$ENABLED_ACLS" == 0 ] && return
add_firewall_rule
gen_include
}
stop() {
del_firewall_rule
[ $(config_t_get global flush_set "0") = "1" ] && {
uci -q delete ${CONFIG}.@global[0].flush_set
uci -q commit ${CONFIG}
flush_ipset
rm -rf /tmp/etc/passwall_tmp/singbox*
rm -rf /tmp/etc/passwall_tmp/dnsmasq*
}
flush_include
}
arg1=$1
shift
case $arg1 in
RULE_LAST_INDEX)
RULE_LAST_INDEX "$@"
;;
insert_rule_before)
insert_rule_before "$@"
;;
insert_rule_after)
insert_rule_after "$@"
;;
get_ipt_bin)
get_ipt_bin
;;
get_ip6t_bin)
get_ip6t_bin
;;
get_wan_ip)
get_wan_ip
;;
get_wan6_ip)
get_wan6_ip
;;
filter_direct_node_list)
filter_direct_node_list
;;
stop)
stop
;;
start)
start
;;
*) ;;
esac
|
294coder/Efficient-MIF | 6,593 | utils/save_checker.py | # author: Zihan
# date: 2023/11/12
from dataclasses import dataclass
import numpy as np
from pathlib import Path
import shutil
from functools import partial
import sys
sys.path.append('./')
from utils.log_utils import easy_logger
from utils.misc import is_none
logger = easy_logger(func_name='save_checker')
@dataclass()
class BestMetricSaveChecker:
_best_metric: "float | str"
metric_name: "str | list[str]"
check_order: str
def __init__(self,
metric_name: "str | list[str] | dict[str, float]"=None,
check_order: str=None,
*,
avg_metrics_weights: dict[str, float]=None):
# overide the avg_metrics_weights if it is None
if is_none(avg_metrics_weights):
if isinstance(metric_name, (list, tuple)):
self.metric_name = metric_name
avg_metrics_weights = {m: 1. for m in metric_name}
elif isinstance(metric_name, dict):
self.metric_name = list(metric_name.keys())
avg_metrics_weights = metric_name
elif isinstance(metric_name, str):
self.metric_name = metric_name
avg_metrics_weights = {metric_name: 1.}
else:
raise ValueError(f'@metric_name should be a str or a list of str or a dict of str->float but got {metric_name}')
else:
self.metric_name = metric_name
if is_none(metric_name):
logger.warning(f'No metric name provided, and SaveChecker will alway return True')
assert check_order in ['up', 'down', None]
if check_order is None:
check_order = self._default_setting()
self.check_order = check_order
self.avg_metrics_weights = avg_metrics_weights
self.avg_metrics = isinstance(self.metric_name, list) and avg_metrics_weights is not None
logger.info(f'SaveChecker initialized with metric_name={self.metric_name}, ',
f'check_order={check_order}, avg_metrics_weights={avg_metrics_weights}')
if isinstance(metric_name, list):
assert not is_none(avg_metrics_weights), \
f'@avg_metrics_weights should not be None when @metric_name is a list'
for m in metric_name:
assert m in avg_metrics_weights.keys(), \
f'@avg_metrics_weights should have key {m} but got {avg_metrics_weights}'
if check_order == 'up':
default_best_metric_val = -np.Inf
elif check_order == 'down':
default_best_metric_val = np.Inf
else:
default_best_metric_val = 'none'
self._best_metric = default_best_metric_val
self._check_fn = (lambda new, old: new > old) if check_order=='up' else \
(lambda new, old: new <= old)
def _default_setting(self):
metric_name = self.metric_name.lower()
if is_none(metric_name):
return 'none'
_default_dict = {
'psnr': 'up',
'ssim': 'up',
'sam': 'down',
'ergas': 'down',
'cc': 'up',
'scc': 'up',
'sd': 'up',
'en': 'up',
'sf': 'up',
'vif': 'up',
'mse': 'down',
'ag': 'up',
'mi': 'up',
}
if check_order := _default_dict.get(metric_name):
if check_order is None:
raise ValueError(f'No default setting for metric {metric_name}, ' +
'you should provide @check_order manually')
else:
return check_order
def __call__(self, val_metrics: dict[str, float], *args):
if is_none(self.metric_name):
return True
if self.avg_metrics:
for m in self.metric_name:
assert m in val_metrics.keys(), f'@val_metrics should have key {m} but got {val_metrics}'
else:
assert self.metric_name in val_metrics.keys(), f'@val_metrics should have key {self.metric_name} but got {val_metrics}'
if not self.avg_metrics:
new_val = val_metrics[self.metric_name]
else:
new_val = np.sum([val_metrics[k] * self.avg_metrics_weights[k] for k in self.metric_name])
prev_val = self._best_metric
_save = self._check_fn(new_val, prev_val)
if _save:
self._best_metric = new_val
return _save
@property
def best_metric(self):
return self._best_metric
# TODO: need test
class SavedWeightsNumMonitor:
def __init__(self,
saved_weight_path: str,
glob_pattern: str='ep*/',
sort_saved_fn: callable=None,
kept_num: int=4,
rm_verbose: bool=False):
self.saved_weight_path = Path(saved_weight_path)
self.glob_pattern = glob_pattern
self.sort_saved_fn = sort_saved_fn
self.kept_num = kept_num
self.rm_verbose = rm_verbose
if rm_verbose:
self.logger = easy_logger()
@property
def saved_weight_path(self):
return self.saved_weight_path.glob(self.glob_pattern)
@property
def saved_num(self):
# excluse .pth in the base dir
return len(list(self.saved_weight_path))
def rm_for_keep_k(self):
if self.saved_num > self.kept_num:
saved_name = list(map(lambda x: x.name, self.saved_weight_path))
sort_fn = self.sort_saved_fn if self.sort_saved_fn is not None else \
partial(sorted, key=lambda x: int(x))
sorted_saved_name = sort_fn(saved_name)
rm_saved_name = sorted_saved_name[self.kept_num:]
for p in rm_saved_name:
shutil.rmtree(self.saved_weight_path / p)
if self.rm_verbose:
self.logger.print(f'remove {p}')
def __call__(self):
self.rm_for_keep_k()
if __name__ == '__main__':
checker = BestMetricSaveChecker({"psnr": 0.1, "ssim": 0.2}, 'up')
val_d1 = {'sam': 2.3, 'psnr': 10, 'ssim':0.8}
val_d2 = {'sam': 2.4, 'psnr': 12, 'ssim':0.9}
print(checker.best_metric)
print(checker.metric_name)
print(checker(val_d1))
print(checker.best_metric)
print(checker(val_d2))
print(checker.best_metric) |
294coder/Efficient-MIF | 2,092 | utils/model_perf_utils.py | import torch
import time
import pynvml
from tqdm import trange
def get_gpu_memory_usage(device_idx: int):
pynvml.nvmlInit()
handle = pynvml.nvmlDeviceGetHandleByIndex(device_idx)
info = pynvml.nvmlDeviceGetMemoryInfo(handle)
print(f"Total GPU memory: {info.total / 1024**3:.2f} GB")
print(f"Used GPU memory: {info.used / 1024**3:.2f} GB")
print(f"Free GPU memory: {info.free / 1024**3:.2f} GB")
pynvml.nvmlShutdown()
def measure_throughput(model: "torch.nn.Module",
input_size: "list[tuple[int, ...]] | tuple[int, ...] | None",
batch_size: int,
num_warmup: int=10,
num_iterations: int=50):
device = next(model.parameters()).device
device_idx = int(str(device)[-1])
get_gpu_memory_usage(device_idx)
model.eval()
if isinstance(input_size[0], tuple):
dummy_input = []
for inp_shape in input_size:
if inp_shape is None:
inp = torch.randn(batch_size, device=device)
else:
inp = torch.randn(batch_size, *inp_shape, device=device)
dummy_input.append(inp)
elif isinstance(input_size, (list, tuple)):
dummy_input = [torch.randn(batch_size, *input_size, device=device)]
elif input_size is None:
dummy_input = [torch.randn(batch_size, device=device)]
print(f"Warming up {num_warmup} iterations ...")
with torch.no_grad():
for _ in trange(num_warmup):
_ = model(*dummy_input)
torch.cuda.synchronize()
print(f"Measuring model for {num_iterations} iterations ...")
total_time = 0
with torch.no_grad():
for _ in trange(num_iterations):
start_time = time.time()
_ = model(*dummy_input)
torch.cuda.synchronize()
end_time = time.time()
total_time += (end_time - start_time)
images_per_second = (num_iterations * batch_size) / total_time
print(f"Throughput: {images_per_second:.2f} images/second")
return images_per_second
|
294coder/Efficient-MIF | 3,012 | utils/print_helper.py | #
# For licensing see accompanying LICENSE file.
# Copyright (C) 2022 Apple Inc. All Rights Reserved.
#
# only for printing in color
import time
from typing import Optional
import sys
import os
text_colors = {
"logs": "\033[34m", # 033 is the escape code and 34 is the color code
"info": "\033[32m",
"warning": "\033[33m",
"debug": "\033[93m",
"error": "\033[31m",
"bold": "\033[1m",
"end_color": "\033[0m",
"light_red": "\033[36m",
}
def get_curr_time_stamp() -> str:
return time.strftime("%Y-%m-%d %H:%M:%S")
def error(message: str) -> None:
time_stamp = get_curr_time_stamp()
error_str = (
text_colors["error"]
+ text_colors["bold"]
+ "ERROR "
+ text_colors["end_color"]
)
print("{} - {} - {}".format(time_stamp, error_str, message), flush=True)
print("{} - {} - {}".format(time_stamp, error_str, "Exiting!!!"), flush=True)
exit(-1)
def color_text(in_text: str) -> str:
return text_colors["light_red"] + in_text + text_colors["end_color"]
def log(message: str) -> None:
time_stamp = get_curr_time_stamp()
log_str = (
text_colors["logs"] + text_colors["bold"] + "LOGS " + text_colors["end_color"]
)
print("{} - {} - {}".format(time_stamp, log_str, message))
def warning(message: str) -> None:
time_stamp = get_curr_time_stamp()
warn_str = (
text_colors["warning"]
+ text_colors["bold"]
+ "WARNING"
+ text_colors["end_color"]
)
print("{} - {} - {}".format(time_stamp, warn_str, message))
def info(message: str, print_line: Optional[bool] = False) -> None:
time_stamp = get_curr_time_stamp()
info_str = (
text_colors["info"] + text_colors["bold"] + "INFO " + text_colors["end_color"]
)
print("{} - {} - {}".format(time_stamp, info_str, message))
if print_line:
double_dash_line(dashes=150)
def debug(message: str) -> None:
time_stamp = get_curr_time_stamp()
log_str = (
text_colors["debug"]
+ text_colors["bold"]
+ "DEBUG "
+ text_colors["end_color"]
)
print("{} - {} - {}".format(time_stamp, log_str, message))
def double_dash_line(dashes: Optional[int] = 75) -> None:
print(text_colors["error"] + "=" * dashes + text_colors["end_color"])
def singe_dash_line(dashes: Optional[int] = 67) -> None:
print("-" * dashes)
def print_header(header: str) -> None:
double_dash_line()
print(
text_colors["info"]
+ text_colors["bold"]
+ "=" * 50
+ str(header)
+ text_colors["end_color"]
)
double_dash_line()
def print_header_minor(header: str) -> None:
print(
text_colors["warning"]
+ text_colors["bold"]
+ "=" * 25
+ str(header)
+ text_colors["end_color"]
)
def disable_printing():
sys.stdout = open(os.devnull, "w")
def enable_printing():
sys.stdout = sys.__stdout__
if __name__ == '__main__':
log('12345')
|
294coder/Efficient-MIF | 2,118 | utils/save_to_mat.py | import h5py
import numpy as np
import os
import os.path as osp
import glob
import scipy.io as io
from tqdm import tqdm
import matplotlib.pyplot as plt
def process_mat_to_single_mat(file, show=True, force_all=True):
name = file.split('/')[-1].strip('.mat')
path = f'/Data2/ZiHanCao/exps/panformer/visualized_img/{name}'
print(path)
save_prefix = 'output_mulExm_'
if not osp.exists(path) or force_all:
os.makedirs(path, exist_ok=True)
print(f'make dir {name}')
else:
print('already process this dir')
return
mat_file = io.loadmat(file)
print(f'has keys: {mat_file.keys()}')
if 'wv3' in path:
rgb_index = [4,2,0]
const = 2047
elif 'gf5' in path:
rgb_index = [49, 39, 19]
const = 1
elif 'gf' in path or 'qb' in path:
rgb_index = [2,1,0]
const = 1023
elif 'cave' in path or 'harvard'in path:
rgb_index = [29,19,9]
const = 1
elif 'houston' in path:
rgb_index = [39, 29, 19]
const = 1
else:
rgb_index = [29, 19, 9]
const = 1
sr = mat_file.get('sr')
if show:
ncols = int(np.ceil(sr.shape[0]/4))
fig, axes = plt.subplots(4, ncols, figsize=(ncols*4, 4*4))
axes = axes.flatten()
if sr is None:
print('has no key sr')
else:
bar = tqdm(range(sr.shape[0]))
for i in bar:
save_path = osp.join(path, save_prefix+f'{i}.mat')
sr_i = np.transpose(sr[i, ...], [1,2,0])
save_d = {'sr': sr_i}
io.savemat(save_path, save_d)
bar.set_description(f'save {i}.mat')
if show:
axes[i].imshow(sr_i[..., rgb_index] / const)
axes[i].set_axis_off()
if show:
plt.tight_layout()
fig.savefig(path + '/sr.png')
print('---------------'*10)
files = glob.glob('/Data2/ZiHanCao/exps/panformer/visualized_img/*.mat')
print('found files: ', files)
for file in files:
process_mat_to_single_mat(file, force_all=True)
|
2977094657/DidaAPI | 972 | frontend/docs/api/user/get-user-info.md | # 获取用户信息
获取当前登录用户的基本信息。
## 接口信息
- **原始API**: `https://api.dida365.com/api/v2/user/profile`
- **方法**: `GET`
- **认证**: 需要登录认证
## 请求参数
无需参数,使用当前认证会话。
## 响应格式
### 成功响应
```json
{
"username": "string",
"email": "string",
"picture": "string",
"locale": "zh_CN",
"userCode": "string",
"timezone": "Asia/Shanghai",
"nickname": "string",
"phone": "string",
"createdTime": "string",
"lastLoginTime": "string",
"premium": false,
"subscriptionType": "string",
"subscriptionExpiry": "string"
}
```
## 响应字段说明
| 字段 | 类型 | 描述 |
|------|------|------|
| username | string | 用户名 |
| email | string | 邮箱地址 |
| picture | string | 头像URL |
| locale | string | 语言设置 |
| userCode | string | 用户代码 |
| timezone | string | 时区设置 |
| nickname | string | 昵称 |
| phone | string | 手机号 |
| createdTime | string | 注册时间 |
| lastLoginTime | string | 最后登录时间 |
| premium | boolean | 是否为高级用户 |
| subscriptionType | string | 订阅类型 |
| subscriptionExpiry | string | 订阅到期时间 |
|
2977094657/DidaAPI | 2,818 | frontend/docs/api/custom/export-tasks-excel.md | # 导出任务到Excel
将用户的所有任务导出为Excel文件,包含全部任务、已完成任务、放弃任务、垃圾桶任务四个工作表。
## 接口信息
- **接口URL**: `http://localhost:8000/custom/export/tasks/excel`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 本项目自定义接口
## 功能说明
此接口会将用户的所有任务数据导出为Excel文件,包含以下四个工作表:
### 工作表1:全部任务
包含当前所有任务(未完成和已完成),数据来源于 `/api/v2/batch/check/0` 接口。
### 工作表2:已完成任务
包含历史已完成的任务,数据来源于 `/api/v2/project/all/closed?status=Completed` 接口。
**分页获取机制**:
- 第一次请求不携带`to`参数
- 后续请求使用上次响应最后一个任务的`completedTime`作为`to`参数
- 重复请求直到返回数据少于50条为止,确保获取所有历史数据
### 工作表3:放弃任务
包含历史放弃的任务,数据来源于 `/api/v2/project/all/closed?status=Abandoned` 接口。
**分页获取机制**:
- 第一次请求不携带`to`参数
- 后续请求使用上次响应最后一个任务的`completedTime`作为`to`参数
- 重复请求直到返回数据少于50条为止,确保获取所有历史数据
### 工作表4:垃圾桶任务
包含已删除的任务,数据来源于 `/api/v2/project/all/trash/page` 接口。
## 导出字段
每个工作表包含任务的完整字段信息(展平后),包括但不限于:
### 基本信息
- 任务ID
- 任务标题
- 任务内容
- 任务描述
- 项目ID
- 项目名称
- 排序顺序
### 状态信息
- 任务状态(文本描述)
- 状态代码(数字)
- 优先级
- 完成进度
- 删除状态
### 时间信息
- 创建时间
- 修改时间
- 开始日期
- 截止日期
- 置顶时间
- 完成时间
- 删除时间
### 时区和时间设置
- 时区
- 是否浮动时间
- 是否全天任务
### 重复设置
- 重复任务ID
- 重复标志
- 重复来源
- 首次重复日期
### 提醒设置
- 提醒设置
- 提醒列表
- 排除日期
### 层级关系
- 父任务ID
- 子任务ID列表
### 其他属性
- 标签列表
- 子项目
- 附件数量
- 评论数量
- 列ID
- 类型
- 图片模式
- 创建者ID
- 删除者ID
- 实体标签
- 番茄钟摘要
- 专注摘要
- 附件详情
## 请求示例
```http
GET http://localhost:8000/custom/export/tasks/excel HTTP/1.1
Host: localhost:8000
Accept: application/vnd.openxmlformats-officedocument.spreadsheetml.sheet
```
## 响应格式
### 成功响应
**状态码**: `200 OK`
**Content-Type**: `application/vnd.openxmlformats-officedocument.spreadsheetml.sheet`
**响应头**:
```http
Content-Disposition: attachment; filename=滴答清单任务导出_20250106_143022.xlsx
```
**响应体**: Excel文件的二进制数据
### 错误响应
**状态码**: `401 Unauthorized`
```json
{
"detail": "未设置认证会话,请先完成登录"
}
```
**状态码**: `500 Internal Server Error`
```json
{
"detail": "导出失败: 具体错误信息"
}
```
## 使用说明
1. **认证要求**: 必须先完成认证,设置有效的会话
2. **文件格式**: 导出的Excel文件包含四个工作表
3. **文件命名**: 文件名格式为 `滴答清单任务导出_YYYYMMDD_HHMMSS.xlsx`
4. **数据完整性**: 包含任务的所有字段,无遗漏
5. **分页处理**: 自动处理已完成任务和放弃任务的分页,获取所有历史数据
## 辅助接口
### 获取导出信息
**接口URL**: `http://localhost:8000/custom/export/tasks/excel/info`
**请求方法**: `GET`
**功能**: 获取当前用户任务的统计信息,用于导出前预览
**响应示例**:
```json
{
"auth_status": true,
"all_tasks_count": 150,
"completed_tasks_count": 89,
"abandoned_tasks_count": 15,
"trash_tasks_count": 12,
"session_info": {
"has_session": true,
"session_id": "session_123",
"created_at": "2025-01-06T14:30:22"
}
}
```
## 注意事项
1. 导出过程可能需要一些时间,特别是当任务数量较多时
2. 已完成任务和放弃任务会通过分页获取所有历史数据,确保数据完整性
3. 如果某个数据源获取失败,对应的工作表将为空,但不影响其他工作表
4. 建议在网络状况良好时进行导出操作
5. 导出的Excel文件可以用Microsoft Excel、WPS Office等软件打开
## 相关接口
- [获取所有任务](../tasks/get-all-tasks.md) - 全部任务数据源
- [获取已完成任务](../tasks/get-completed-tasks.md) - 已完成任务数据源
- [获取已完成/已放弃任务](../tasks/get-completed-tasks.md) - 放弃任务数据源(使用status=Abandoned参数)
- [获取垃圾桶任务](../tasks/get-trash-tasks.md) - 垃圾桶任务数据源
|
2977094657/DidaAPI | 3,103 | frontend/docs/api/custom/export-focus-excel.md | # 导出专注记录到Excel
将用户的所有专注记录导出为Excel文件,包含完整的专注时间线数据。
## 接口信息
- **接口URL**: `http://localhost:8000/custom/export/focus/excel`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 本项目自定义接口
## 功能说明
此接口会将用户的所有专注记录数据导出为Excel文件,包含以下工作表:
### 工作表:专注记录时间线
包含所有专注记录的详细信息,数据来源于 `/api/v2/pomodoros/timeline` 接口,支持自动分页获取所有历史数据。
## 导出字段
采用紧凑型展示方式,每行一个专注会话,通过特殊格式直观展示专注和暂停模式:
### 主要字段
- **会话ID**: 专注会话的唯一标识符
- **会话时间**: 会话的日期和时间范围,格式:`2025-06-05 11:44 - 17:33`
- **总时长**: 会话总时长,格式:`5小时49分钟`
- **暂停时长**: 会话中的暂停时长,格式:`1小时11分钟`
- **任务标题**: 专注的任务标题(多个任务用分号分隔)
- **项目**: 任务所属项目(多个项目用分号分隔)
### 核心展示字段
- **专注时间段**: 详细的专注时间轴,格式示例:
```
11:44-12:09(25分钟) → [暂停1小时11分钟] → 13:20-17:33(4小时13分钟)
```
- **暂停模式**: 暂停情况的统计分析,格式示例:
```
暂停1次(1小时11分钟)
暂停3次(总计2小时30分钟, 平均50分钟)
无暂停
```
### 分析字段
- **效率(%)**: 实际专注时长占总时长的百分比
- **时间段数量**: 包含的专注时间段数量
- **会话类型**: 会话类型标识
- **实体标签**: 版本控制标签
### 展示示例
| 会话ID | 会话时间 | 总时长 | 暂停时长 | 任务标题 | 专注时间段 | 暂停模式 | 效率(%) |
|--------|----------|--------|----------|----------|------------|----------|---------|
| 6841831a... | 2025-06-05 11:44 - 17:33 | 5小时49分钟 | 1小时11分钟 | 从web端抓包所有api | 11:44-12:09(25分钟) → [暂停1小时11分钟] → 13:20-17:33(4小时13分钟) | 暂停1次(1小时11分钟) | 79.7% |
| 6841599c... | 2025-06-05 08:47 - 10:11 | 1小时24分钟 | 0分钟 | 从web端抓包所有api | 08:47-10:11(1小时24分钟) | 无暂停 | 100.0% |
**重要说明**:
- 每个专注会话只占一行,信息紧凑但完整
- "专注时间段"字段能直观显示专注过程中的暂停和继续情况
- 时间格式自动优化显示(秒/分钟/小时)
- 暂停模式提供统计分析,便于了解专注习惯
## 请求示例
```http
GET http://localhost:8000/custom/export/focus/excel HTTP/1.1
Host: localhost:8000
Accept: application/vnd.openxmlformats-officedocument.spreadsheetml.sheet
```
## 响应格式
### 成功响应
**状态码**: `200 OK`
**Content-Type**: `application/vnd.openxmlformats-officedocument.spreadsheetml.sheet`
**响应头**:
```http
Content-Disposition: attachment; filename=滴答清单专注记录导出_20250106_143022.xlsx
```
**响应体**: Excel文件的二进制数据
### 错误响应
**状态码**: `401 Unauthorized`
```json
{
"detail": "未设置认证会话,请先完成登录"
}
```
**状态码**: `500 Internal Server Error`
```json
{
"detail": "导出失败: 具体错误信息"
}
```
## 分页处理说明
专注记录时间线接口支持分页,本导出功能会自动处理分页:
1. **分页参数**: 使用 `to` 参数进行分页,值为上一页最后一条记录的 `startTime`
2. **时间转换**: 自动将时间字符串转换为中国时间戳(+8小时)
3. **自动分页**: 持续获取直到没有更多数据
4. **数据完整性**: 确保获取所有历史专注记录
## 使用说明
1. **认证要求**: 必须先完成认证,设置有效的会话
2. **文件格式**: 导出的Excel文件包含专注记录时间线工作表
3. **文件命名**: 文件名格式为 `滴答清单专注记录导出_YYYYMMDD_HHMMSS.xlsx`
4. **数据完整性**: 包含专注记录的所有字段,无遗漏
5. **分页处理**: 自动处理分页,获取所有历史数据
## 辅助接口
### 获取导出信息
**接口URL**: `http://localhost:8000/custom/export/focus/excel/info`
**请求方法**: `GET`
**功能**: 获取当前用户专注记录的统计信息,用于导出前预览
**响应示例**:
```json
{
"auth_status": true,
"focus_records_count_estimate": "31+ (需要分页获取完整数据)",
"session_info": {
"has_session": true,
"session_id": "session_123",
"created_at": "2025-01-06T14:30:22"
},
"note": "专注记录数量为预估值,实际导出时会获取所有历史数据"
}
```
## 注意事项
1. 导出过程可能需要一些时间,特别是当专注记录数量较多时
2. 会通过分页自动获取所有历史专注记录,确保数据完整性
3. 如果数据源获取失败,导出将失败并返回错误信息
4. 建议在网络状况良好时进行导出操作
5. 导出的Excel文件可以用Microsoft Excel、WPS Office等软件打开
## 相关接口
- [获取专注记录时间线](../pomodoros/focus-timeline.md) - 专注记录数据源
- [获取专注详情分布](../pomodoros/focus-distribution.md) - 专注分布统计
- [导出任务到Excel](./export-tasks-excel.md) - 任务导出功能
|
281677160/openwrt-package | 3,222 | luci-app-passwall/root/usr/share/passwall/tasks.sh | #!/bin/sh
## 循环更新脚本
CONFIG=passwall
APP_PATH=/usr/share/$CONFIG
TMP_PATH=/tmp/etc/$CONFIG
LOCK_FILE=/tmp/lock/${CONFIG}_tasks.lock
CFG_UPDATE_INT=0
config_n_get() {
local ret=$(uci -q get "${CONFIG}.${1}.${2}" 2>/dev/null)
echo "${ret:=$3}"
}
config_t_get() {
local index=${4:-0}
local ret=$(uci -q get "${CONFIG}.@${1}[${index}].${2}" 2>/dev/null)
echo "${ret:=${3}}"
}
exec 99>"$LOCK_FILE"
flock -n 99
if [ "$?" != 0 ]; then
exit 0
fi
while true
do
if [ "$CFG_UPDATE_INT" -ne 0 ]; then
stop_week_mode=$(config_t_get global_delay stop_week_mode)
stop_interval_mode=$(config_t_get global_delay stop_interval_mode)
stop_interval_mode=$(expr "$stop_interval_mode" \* 60)
if [ -n "$stop_week_mode" ]; then
[ "$stop_week_mode" = "8" ] && {
[ "$(expr "$CFG_UPDATE_INT" % "$stop_interval_mode")" -eq 0 ] && /etc/init.d/$CONFIG stop > /dev/null 2>&1 &
}
fi
start_week_mode=$(config_t_get global_delay start_week_mode)
start_interval_mode=$(config_t_get global_delay start_interval_mode)
start_interval_mode=$(expr "$start_interval_mode" \* 60)
if [ -n "$start_week_mode" ]; then
[ "$start_week_mode" = "8" ] && {
[ "$(expr "$CFG_UPDATE_INT" % "$start_interval_mode")" -eq 0 ] && /etc/init.d/$CONFIG start > /dev/null 2>&1 &
}
fi
restart_week_mode=$(config_t_get global_delay restart_week_mode)
restart_interval_mode=$(config_t_get global_delay restart_interval_mode)
restart_interval_mode=$(expr "$restart_interval_mode" \* 60)
if [ -n "$restart_week_mode" ]; then
[ "$restart_week_mode" = "8" ] && {
[ "$(expr "$CFG_UPDATE_INT" % "$restart_interval_mode")" -eq 0 ] && /etc/init.d/$CONFIG restart > /dev/null 2>&1 &
}
fi
autoupdate=$(config_t_get global_rules auto_update)
weekupdate=$(config_t_get global_rules week_update)
hourupdate=$(config_t_get global_rules interval_update)
hourupdate=$(expr "$hourupdate" \* 60)
if [ "$autoupdate" = "1" ]; then
[ "$weekupdate" = "8" ] && {
[ "$(expr "$CFG_UPDATE_INT" % "$hourupdate")" -eq 0 ] && lua $APP_PATH/rule_update.lua log all cron > /dev/null 2>&1 &
}
fi
TMP_SUB_PATH=$TMP_PATH/sub_tasks
mkdir -p $TMP_SUB_PATH
for item in $(uci show ${CONFIG} | grep "=subscribe_list" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do
if [ "$(config_n_get $item auto_update 0)" = "1" ]; then
cfgid=$(uci show ${CONFIG}.$item | head -n 1 | cut -d '.' -sf 2 | cut -d '=' -sf 1)
remark=$(config_n_get $item remark)
week_update=$(config_n_get $item week_update)
hour_update=$(config_n_get $item interval_update)
echo "$cfgid" >> $TMP_SUB_PATH/${week_update}_${hour_update}
fi
done
[ -d "${TMP_SUB_PATH}" ] && {
for name in $(ls ${TMP_SUB_PATH}); do
week_update=$(echo $name | awk -F '_' '{print $1}')
hour_update=$(echo $name | awk -F '_' '{print $2}')
hour_update=$(expr "$hour_update" \* 60)
cfgids=$(echo -n $(cat ${TMP_SUB_PATH}/${name}) | sed 's# #,#g')
[ "$week_update" = "8" ] && {
[ "$(expr "$CFG_UPDATE_INT" % "$hour_update")" -eq 0 ] && lua $APP_PATH/subscribe.lua start $cfgids cron > /dev/null 2>&1 &
}
done
rm -rf $TMP_SUB_PATH
}
fi
CFG_UPDATE_INT=$(expr "$CFG_UPDATE_INT" + 10)
sleep 600
done 2>/dev/null
|
294coder/Efficient-MIF | 10,627 | utils/metric_sharpening.py | import numpy as np
import torch
from skimage.metrics import structural_similarity, peak_signal_noise_ratio
from functools import partial
from warnings import warn
import sys
sys.path.append('./')
from utils.misc import to_numpy, dict_to_str
from utils._metric_legacy import analysis_accu, indexes_evaluation_FS
class NonAnalysis(object):
def __init__(self):
self.acc_ave = {} # only used as attribution
pass
def __call__(self, *args, **kwargs):
pass
def __repr__(self):
return 'NonAnalysis()'
# TODO: need to be tested the new metric analysis
# FIXME: this python code is not same as matlab code, you should use matlab code to get the real accuracy
# only used in training and validate
class AnalysisPanAcc(object):
def __init__(self, ratio=4, ref=True, ergas_ratio: int = 4, **unref_factory_kwargs):
"""pansharpening metric analysis class
Args:
ratio (int, optional): fusion ratio. Defaults to 4.
ref (bool, optional): reduce-resolution or full-resolution. Defaults to True.
ergas_ratio (int, optional): previous api (may decrepated soon). Defaults to 4.
unref_factory_kwargs(dict): sensor, default_max_value. Defaults to {'sensor': 'default', 'default_max_value': None}.
Raises:
ValueError: _description_
"""
# ergas_ratio is decrepated
if ratio is None:
ratio = ergas_ratio
warn('@ergas_ratio is deprecated, use ratio instead')
self.ratio = ratio
self.ref = ref
# metric functions
if ref:
self.__sam_ergas_psnr_cc_one_image = partial(analysis_accu, ratio=ergas_ratio, choices=5)
self.ssim = ssim_batch_tensor_metric
else:
# @sensor in ['QB', 'IKONOS', 'WV2', 'WV3', 'default']
assert 'sensor' in unref_factory_kwargs or 'default_max_value' in unref_factory_kwargs, \
'@sensor or @default_max_value should be specified in unrefactory_kwargs'
sensor = unref_factory_kwargs.pop('sensor', 'default').upper()
if sensor == 'DEFAULT': warn('sensor is not specified, use default sensor type')
self.default_max_value = unref_factory_kwargs.pop('default_max_value', None)
if self.default_max_value is None:
_default_max_value = {'QB': 2047, 'IKONOS': 1023, 'WV2': 2047, 'WV3': 2047,
'GF2': 1023, 'DEFAULT': 2047,
'CAVE_X4': 1, 'CAVE_X8': 1, 'HARVARD_X': 1, 'HARVARD_X8': 1,
'GF5':1, 'GF2-GF5': 1,}
self.default_max_value = _default_max_value.get(sensor)
print(f'>>> @default_max_value is not specified, set it according to @sensor:'
f'{sensor, self.default_max_value}\n', '-'*20)
self.FS_metric_fn = partial(indexes_evaluation_FS, L=11, Qblocks_size=32, sensor=sensor,
th_values=0, ratio=ratio, flagQNR=False)
# tracking accuracy
self._acc_d = {}
self._call_n = 0
self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if ref else \
{'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
@property
def empty_acc(self):
return {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \
{'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
@staticmethod
def permute_dim(*args, permute_dims=(1, 2, 0)):
l = []
for i in args:
l.append(i.permute(*permute_dims))
return l
@staticmethod
def _sum_acc(d_ave, d_now, n, n2=1):
assert len(d_ave) == len(d_now)
for k in d_ave.keys():
v2 = d_now[k] * n2
d_ave[k] *= n
d_ave[k] += v2.cpu().item() if isinstance(v2, torch.Tensor) else v2
return d_ave
@staticmethod
def _average_acc(d_ave, n):
for k in d_ave.keys():
d_ave[k] /= n
return d_ave
def sam_ergas_psnr_cc_batch(self, b_gt, b_pred):
n = b_gt.shape[0]
# input shape should be [B, C, H, W]
acc_ds = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0.}
for i, (img1, img2) in enumerate(zip(b_gt, b_pred)):
img1, img2 = self.permute_dim(img1, img2)
acc_d = self.__sam_ergas_psnr_cc_one_image(img1, img2)
acc_ds = self._sum_acc(acc_ds, acc_d, i)
acc_ds = self._average_acc(acc_ds, i + 1)
return acc_ds
def D_lambda_D_s_HQNR_batch(self, sr=None, ms=None, lms=None, pan=None):
assert sr is not None and lms is not None and pan is not None and ms is not None
if ms is None:
ms = torch.nn.functional.interpolate(lms, scale_factor=1/self.rato, mode='bilinear', align_corners=False)
acc_ds = {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
sr, ms, lms, pan = self.permute_dim(sr, ms, lms, pan, permute_dims=(0, 2, 3, 1))
sr, ms, lms, pan = to_numpy(sr, ms, lms, pan)
_max_value = getattr(self, 'default_max_value')
sr, ms, lms, pan = map(lambda x: np.clip(x * _max_value, 0, _max_value), [sr, ms, lms, pan])
for i, (sr_i, ms_i, lms_i, pan_i) in enumerate(zip(sr, ms, lms, pan)):
QNR_index, D_lambda, D_S, _ = self.FS_metric_fn(I_F=sr_i, I_MS_LR=ms_i, I_MS=lms_i, I_PAN=pan_i)
acc_d = dict(HQNR=QNR_index, D_lambda=D_lambda, D_S=D_S)
acc_ds = self._sum_acc(acc_ds, acc_d, i)
acc_ds = self._average_acc(acc_ds, i + 1)
return acc_ds
def once_batch_call(self, **kwargs):
if self.ref:
acc_d1 = self.sam_ergas_psnr_cc_batch(**kwargs)
acc_ssim = self.ssim(**kwargs)
acc_d1['SSIM'] = acc_ssim
else:
acc_d1 = self.D_lambda_D_s_HQNR_batch(**kwargs)
self._acc_d = acc_d1
return acc_d1
def _call_check_args_to_kwargs(self, *args):
def may_np_to_tensor(d):
for k, v in d.items():
if not isinstance(v, torch.Tensor):
d[k] = torch.tensor(v, dtype=torch.float32)
elif v.dtype != torch.float32:
d[k] = v.float()
return d
if len(args) == 2:
assert self.ref, 'ref mode should have 2 args'
kwargs = dict(b_gt=args[0], b_pred=args[1])
elif len(args) == 3:
assert not self.ref, 'unref mode should have more than 2 args'
kwargs = dict(sr=args[0], lms=args[1], pan=args[2])
elif len(args) == 4:
assert not self.ref, 'unref mode should have more than 2 args'
kwargs = dict(sr=args[0], ms=args[1], lms=args[2], pan=args[3])
else:
raise ValueError('args should have 2 or 4 elements')
return may_np_to_tensor(kwargs)
def __call__(self, *args):
"""
Args:
ref mode:
b_gt (torch.Tensor): [b, c, h, w]
b_pred (torch.Tensor): [b, c, h, w]
unref mode:
sr (torch.Tensor): [b, c, h, w]
ms (torch.Tensor, optional): [b, c, h/ratio, w/ratio]
lms (torch.Tensor): [b, c, h, w]
pan (torch.Tensor): [b, c, h, w]
"""
kwargs = self._call_check_args_to_kwargs(*args)
n = args[0].shape[0]
self.acc_ave = self._sum_acc(self.acc_ave, self.once_batch_call(**kwargs), self._call_n, n2=n)
self.acc_ave = self._average_acc(self.acc_ave, self._call_n + n)
self._call_n += n
return self.acc_ave
def clear_history(self, verbose=False):
if verbose:
print('>> AccAnalysis: clear history')
self._acc_d = {}
self._call_n = 0
self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \
{'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
def print_str(self, decimals=6):
return dict_to_str(self.acc_ave, decimals=decimals)
def result_str(self):
return self.print_str()
def __repr__(self) -> str:
repr_str = f'AnalysisPanAcc(ratio={self.ratio}, ref={self.ref}):'
repr_str += f'\n{self.print_str()}'
return repr_str
def normalize_to_01(x):
# normalize tensor to [0, 1]
if isinstance(x, torch.Tensor):
x -= x.flatten(-2).min(-1, keepdim=True)[0][..., None]
x /= x.flatten(-2).max(-1, keepdim=True)[0][..., None]
elif isinstance(x, np.ndarray):
x -= x.min((-2, -1), keepdims=True)
x /= x.max((-2, -1), keepdims=True)
return x
def psnr_one_img(img_gt, img_test):
"""
calculate PSNR for one image
:param img_gt: ground truth image, numpy array, shape [H, W, C]
:param img_test: test or inference image, numpy array, shape [H, W, C]
:return: PSNR, float type
"""
assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size'
return peak_signal_noise_ratio(img_gt, img_test)
def psnr_batch_tensor_metric(b_gt, b_pred):
"""
calculate PSNR for batch tensor images
:param b_gt: tensor, shape [B, C, H, W]
:param b_test: tensor, shape [B, C, H, W]
:return:
"""
assert b_gt.shape[0] == b_pred.shape[0]
bs = b_gt.shape[0]
psnr = 0.
for gt, t in zip(b_gt, b_pred):
psnr += psnr_one_img(*(to_numpy(gt, t)))
return psnr / bs
def ssim_one_image(img_gt, img_test, channel_axis=0):
assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size'
return structural_similarity(img_gt, img_test, channel_axis=channel_axis,
data_range=1.)
def ssim_batch_tensor_metric(b_gt, b_pred):
assert b_gt.shape[0] == b_pred.shape[0]
bs = b_gt.shape[0]
ssim = 0.
for gt, t in zip(b_gt, b_pred):
ssim += ssim_one_image(*(to_numpy(gt, t)), channel_axis=0)
return ssim / bs
if __name__ == '__main__':
sr = torch.rand(4, 3, 256, 256)
ms = torch.rand(4, 3, 64, 64)
lms = torch.rand(4, 3, 256, 256)
pan = torch.rand(4, 3, 256, 256)
gt = torch.rand(4, 3, 256, 256)
analysis = AnalysisPanAcc(ref=False, ratio=4, default_max_value=2047)
for i in range(2):
analysis(sr[i:i+2], ms[i:i+2], lms[i:i+2], pan[i:i+2])
print(analysis.print_str())
|
294coder/Efficient-MIF | 8,375 | utils/load_params.py | from functools import partial
from numpy import isin
import torch
import logging
from collections import OrderedDict
from safetensors.torch import load_file
from packaging import version
from utils.log_utils import easy_logger
def module_load(path,
model,
device,
ddp_rank=None,
strict=True,
spec_key='shadow_params',
logger=None,
full_unmatched_log=True):
if logger is None:
logger = easy_logger(func_name=__name__)
model = model.to(device if ddp_rank is None else ddp_rank)
place = device if ddp_rank is None else {'cuda:%d' % 0: 'cuda:%d' % ddp_rank}
if isinstance(place, torch.device):
place = str(place)
if path.endswith('pth') or path.endswith('pt'):
if version.parse(torch.__version__) >= version.parse('2.4.0'):
load_engine = partial(torch.load, map_location=place, weights_only=False)
else:
load_engine = partial(torch.load, map_location=place)
elif path.endswith('safetensors'):
load_engine = lambda weight_path, map_location: OrderedDict(load_file(weight_path, device=map_location))
else:
raise ValueError
try:
params = load_engine(path, map_location=place)
except Exception:
# TODO: the exception is not used
logger.print('>>> did not find the pth file, try to find in used_weights/ and ununsed_weights/...', logging.INFO)
try:
path_used = path.replace('weight/', 'weight/used_weights/')
params = load_engine(path_used, map_location=place)
except Exception:
path_ununsed = path.replace('weight/', 'weight/unused_weights/')
params = load_engine(path_ununsed, map_location=place)
# parse key
if spec_key is not None:
parsed_keys = spec_key.split('.')
try:
for k in parsed_keys:
params = params[k]
except KeyError:
logger.warning(f'>>> not found parsed model `{spec_key}`, load the model directly \n \n')
_load_fail_flag = False
params_load = params
# may be tedious but useful and safe to avoid 'module.' prefix caused error
if not strict:
logger.warning('model load strict is False, set it to True if you know what you are doing')
def _iter_params_load_fn(model, params_load, strict):
nonlocal _load_fail_flag
if not isinstance(params_load, (list, tuple)):
param_load_ziped = list(params_load.items()) if isinstance(params_load, dict) else params_load
for (s_name, s_param), (name, param) in zip(param_load_ziped, model.named_parameters()):
saved_shape = tuple(s_param.data.shape)
required_shape = tuple(param.data.shape)
if saved_shape != required_shape:
if strict:
logger.print(
f'param shape unmatched, {name} requires: {required_shape}, but got {s_name}: {saved_shape}',
logging.WARNING
)
if not full_unmatched_log:
logger.print('model load failed! shape of params does not match!', logging.ERROR)
raise RuntimeError('model load failed! shape of params does not match!')
else:
_load_fail_flag = True
continue
else:
logger.print(f'skip the shape mismatched param, param name {name}, '
+ f'current shape {required_shape} but loaded shape {saved_shape}', logging.WARNING)
continue
param.data.copy_(s_param.data)
else:
for s_param, param in zip(params_load, model.parameters()):
required_shape = tuple(param.data.shape)
saved_shape = tuple(s_param.data.shape)
if saved_shape != required_shape:
if strict:
logger.print(
f'param shape unmatched, requires: {required_shape}, but got {saved_shape}',
logging.WARNING
)
if not full_unmatched_log:
logger.print('model load failed! shape of params does not match!', logging.ERROR)
raise RuntimeError('model load failed! shape of params does not match!')
else:
_load_fail_flag = True
continue
else:
logger.print(f'skip the shape mismatched param, current shape {required_shape} but loaded shape {saved_shape}', logging.WARNING)
continue
param.data.copy_(s_param.data)
def _load_fn(model, params_load, strict):
if isinstance(params_load, OrderedDict): # ordered dict
model.load_state_dict(params_load, strict=strict)
else:
_iter_params_load_fn(model, params_load, strict)
_load_fn(model, params_load, strict)
if _load_fail_flag:
raise RuntimeError('model load failed! shape of params does not match!')
# except Exception:
# # data parallel mode will save params with keys' prefix is 'module'.
# odict = {}
# for k, v in params_load.items():
# odict['module.' + k] = v
# params_load[spec_key] = odict
# if 'ema' not in spec_key:
# _load_fn(model, params_load, strict)
# else:
# raise RuntimeError('ema model load failed! shape of params does not match!')
logger.print('load pretrain weights', logging.INFO)
return model
# def resume_load(path,
# model,
# optim,
# lr_scheduler,
# ema_model: ExponentialMovingAverage=None,
# specific_resume_lr: float = None,
# specific_epochs: int = None,
# wd_scheduler=None,
# device='cuda:0',
# ddp_rank=None,
# ddp=False):
# # @specific_resume_lr(warning: not recommended):
# # manually specify learning rate when the lr from last break is too low to update model
# # @specific_epochs(warning: not recommended):
# # manually specify total epochs when resuming training
# model.to(device if ddp_rank is None else ddp_rank)
# # assume saved params always on cuda:0
# params = torch.load(path, map_location=device if ddp_rank is None else {'cuda:%d' % 0: 'cuda:%d' % ddp_rank})
# # NOTE: ddp mode will save params with keys' prefix is 'module'.
# # now I remove the prefix for just one card circumstance but it conflict with ddp mode.
# if ddp:
# odict = OrderedDict()
# for k, v in params['model'].items():
# odict['module.' + k] = v
# params['model'] = odict
# model.load_state_dict(params['model'])
# if ema_model is not None:
# ema_model.load_state_dict(params['ema_model'])
# # NOTE: Pytorch 1.12.0 may cause CUDA error in optimizer reloading. see more at
# # https://github.com/pytorch/pytorch/issues/80809#issuecomment-1175211598
# optim.load_state_dict(params['optim'])
# if specific_resume_lr is not None:
# optim.param_groups[0]['lr'] = specific_resume_lr
# lr_scheduler.load_state_dict(params['lr_scheduler'])
# if specific_epochs is not None:
# # FIXME: only support CosineAnnealing lr_scheduler
# lr_scheduler.T_max = specific_epochs
# resume_ep = params['epochs']
# print(f"last training resume! best metrics are {params['metrics']}")
# # warning: if you change total epochs in the resume run, the lr_scheduler may not update lr
# if wd_scheduler is not None:
# wd_scheduler.load_state_dict(params['wd_scheduler'])
# return model, optim, lr_scheduler, wd_scheduler, resume_ep
# else:
# return model, optim, lr_scheduler, resume_ep
|
2977094657/DidaAPI | 2,234 | frontend/docs/api/statistics/task-statistics.md | # 获取任务统计信息
## 接口信息
- **接口路径**: `GET /statistics/tasks`
- **接口描述**: 获取指定日期范围内的任务统计信息
- **请求方式**: GET
- **认证要求**: 需要先完成微信登录获取认证会话
## 请求参数
### Query参数
| 参数名 | 类型 | 必填 | 描述 | 示例值 |
|--------|------|------|------|--------|
| start_date | string | 是 | 开始日期,格式: YYYYMMDD | `20231201` |
| end_date | string | 是 | 结束日期,格式: YYYYMMDD | `20231207` |
## 请求示例
```bash
curl -X GET "http://localhost:8000/statistics/tasks?start_date=20231201&end_date=20231207"
```
## 响应格式
### 成功响应
```json
{
"overdueCompletedCount": 2,
"onTimeCompletedCount": 15,
"noDateCompletedCount": 8,
"uncompletedCount": 5,
"projectStats": [
{
"projectId": "project123",
"projectName": "工作项目",
"completedCount": 10
},
{
"projectId": "project456",
"projectName": "个人项目",
"completedCount": 5
}
],
"tagStats": [
{
"tagName": "重要",
"completedCount": 8
},
{
"tagName": "紧急",
"completedCount": 4
}
]
}
```
### 错误响应
#### 参数错误
```json
{
"detail": [
{
"loc": ["query", "start_date"],
"msg": "field required",
"type": "value_error.missing"
}
]
}
```
#### 未认证
```json
{
"error": "no_auth_session",
"message": "未设置认证会话,请先完成微信登录"
}
```
#### 服务错误
```json
{
"error": "service_error",
"message": "获取任务统计信息失败,请稍后重试"
}
```
## 响应字段说明
| 字段名 | 类型 | 描述 |
|--------|------|------|
| overdueCompletedCount | integer | 逾期完成任务数 |
| onTimeCompletedCount | integer | 按时完成任务数 |
| noDateCompletedCount | integer | 无日期任务完成数 |
| uncompletedCount | integer | 未完成任务数 |
| projectStats | array | 按项目的完成统计 |
| projectStats[].projectId | string | 项目ID |
| projectStats[].projectName | string | 项目名称 |
| projectStats[].completedCount | integer | 该项目完成任务数 |
| tagStats | array | 按标签的完成统计 |
| tagStats[].tagName | string | 标签名称 |
| tagStats[].completedCount | integer | 该标签完成任务数 |
## 使用说明
1. **认证要求**: 需要先调用微信登录接口获取认证会话
2. **日期格式**: 日期必须使用YYYYMMDD格式
3. **统计维度**: 提供多维度的任务完成统计
4. **时间范围**: 支持自定义时间范围查询
## 相关接口
- [获取通用统计信息](./general-statistics.md)
- [获取用户排名统计](../statistics.md)
- [获取任务统计](../tasks/get-tasks-summary.md)
- [微信登录流程](../auth/wechat-login-flow.md)
## 注意事项
- 需要先完成微信登录获取认证会话
- 日期格式必须正确,否则会返回参数错误
- 统计数据基于指定时间范围内的任务
- 项目和标签统计按完成数量降序排列
|
2977094657/DidaAPI | 552 | frontend/docs/api/statistics/index.md | # 统计分析接口
本模块提供了滴答清单的各种统计分析功能,帮助用户了解自己的任务完成情况和使用习惯。
## 接口列表
### 用户排名统计
- [获取用户排名统计](../statistics.md) - 获取用户在滴答清单中的排名和基本统计信息
### 通用统计
- [获取通用统计信息](./general-statistics.md) - 获取概览、成就值、趋势等通用统计信息
### 任务统计
- [获取任务统计信息](./task-statistics.md) - 获取指定日期范围内的任务统计信息
## 使用说明
1. **认证要求**: 所有统计接口都需要先完成微信登录获取认证会话
2. **数据实时性**: 统计数据实时更新,反映最新的用户活动
3. **时间范围**: 部分接口支持自定义时间范围查询
4. **多维度统计**: 提供任务、项目、标签等多维度的统计分析
## 相关模块
- [任务管理](../tasks/get-all-tasks.md) - 任务相关操作
- [习惯管理](../habits.md) - 习惯打卡统计
- [番茄专注](../pomodoros.md) - 专注时间统计
- [认证相关](../auth/) - 登录认证流程
|
281677160/openwrt-package | 23,981 | luci-app-passwall/root/usr/share/passwall/helper_smartdns_add.lua | require "luci.sys"
local api = require "luci.passwall.api"
local appname = "passwall"
local var = api.get_args(arg)
local FLAG = var["-FLAG"]
local SMARTDNS_CONF = var["-SMARTDNS_CONF"]
local LOCAL_GROUP = var["-LOCAL_GROUP"]
local REMOTE_GROUP = var["-REMOTE_GROUP"]
local REMOTE_PROXY_SERVER = var["-REMOTE_PROXY_SERVER"]
local USE_DEFAULT_DNS = var["-USE_DEFAULT_DNS"]
local REMOTE_DNS = var["-REMOTE_DNS"]
local TUN_DNS = var["-TUN_DNS"]
local DNS_MODE = var["-DNS_MODE"]
local REMOTE_FAKEDNS = var["-REMOTE_FAKEDNS"]
local TCP_NODE = var["-TCP_NODE"]
local USE_DIRECT_LIST = var["-USE_DIRECT_LIST"]
local USE_PROXY_LIST = var["-USE_PROXY_LIST"]
local USE_BLOCK_LIST = var["-USE_BLOCK_LIST"]
local USE_GFW_LIST = var["-USE_GFW_LIST"]
local CHN_LIST = var["-CHN_LIST"]
local DEFAULT_PROXY_MODE = var["-DEFAULT_PROXY_MODE"]
local NO_PROXY_IPV6 = var["-NO_PROXY_IPV6"]
local NO_LOGIC_LOG = var["-NO_LOGIC_LOG"]
local NFTFLAG = var["-NFTFLAG"]
local SUBNET = var["-SUBNET"]
local uci = api.uci
local sys = api.sys
local fs = api.fs
local datatypes = api.datatypes
local TMP_PATH = "/tmp/etc/" .. appname
local TMP_ACL_PATH = TMP_PATH .. "/acl"
local RULES_PATH = "/usr/share/" .. appname .. "/rules"
local FLAG_PATH = TMP_ACL_PATH .. "/" .. FLAG
local TMP_CONF_FILE = FLAG_PATH .. "/smartdns.conf"
local config_lines = {}
local tmp_lines = {}
local USE_GEOVIEW = uci:get(appname, "@global_rules[0]", "enable_geoview")
local function log(...)
if NO_LOGIC_LOG == "1" then
return
end
api.log(...)
end
local function is_file_nonzero(path)
if path and #path > 1 then
if sys.exec('[ -s "%s" ] && echo -n 1' % path) == "1" then
return true
end
end
return nil
end
local function insert_unique(dest_table, value, lookup_table)
if not lookup_table[value] then
table.insert(dest_table, value)
lookup_table[value] = true
end
end
local function merge_array(array1, array2)
for i, line in ipairs(array2) do
table.insert(array1, #array1 + 1, line)
end
end
local function insert_array_before(array1, array2, target) --将array2插入到array1的target前面,target不存在则追加
for i, line in ipairs(array1) do
if line == target then
for j = #array2, 1, -1 do
table.insert(array1, i, array2[j])
end
return
end
end
merge_array(array1, array2)
end
local function insert_array_after(array1, array2, target) --将array2插入到array1的target后面,target不存在则追加
for i, line in ipairs(array1) do
if line == target then
for j = 1, #array2 do
table.insert(array1, i + j, array2[j])
end
return
end
end
merge_array(array1, array2)
end
local function get_geosite(list_arg, out_path)
local geosite_path = uci:get(appname, "@global_rules[0]", "v2ray_location_asset") or "/usr/share/v2ray/"
geosite_path = geosite_path:match("^(.*)/") .. "/geosite.dat"
if not is_file_nonzero(geosite_path) then return 1 end
if api.is_finded("geoview") and list_arg and out_path then
sys.exec("geoview -type geosite -append=true -input " .. geosite_path .. " -list '" .. list_arg .. "' -output " .. out_path)
return 0
end
return 1
end
if not fs.access(FLAG_PATH) then
fs.mkdir(FLAG_PATH)
end
local LOCAL_EXTEND_ARG = ""
if LOCAL_GROUP == "nil" then
LOCAL_GROUP = nil
log(" * 注意:国内分组名未设置,可能会导致 DNS 分流错误!")
else
--从smartdns配置中读取参数
local custom_conf_path = "/etc/smartdns/custom.conf"
local options = {
{key = "dualstack_ip_selection", config_key = "dualstack-ip-selection", yes_no = true, arg_yes = "-d yes", arg_no = "-d no", default = "yes"},
{key = "speed_check_mode", config_key = "speed-check-mode", prefix = "-c ", default = "ping,tcp:80,tcp:443"},
{key = "serve_expired", config_key = "serve-expired", yes_no = true, arg_yes = "", arg_no = "-no-serve-expired", default = "yes"},
{key = "response_mode", config_key = "response-mode", prefix = "-r ", default = "first-ping"},
{key = "rr_ttl", config_key = "rr-ttl", prefix = "-rr-ttl "},
{key = "rr_ttl_min", config_key = "rr-ttl-min", prefix = "-rr-ttl-min "},
{key = "rr_ttl_max", config_key = "rr-ttl-max", prefix = "-rr-ttl-max "}
}
-- 从 custom.conf 中读取值,以最后出现的值为准
local custom_config = {}
local f_in = io.open(custom_conf_path, "r")
if f_in then
for line in f_in:lines() do
line = api.trim(line)
if line ~= "" and not line:match("^#") then
local param, value = line:match("^(%S+)%s+(%S+)$")
if param and value then custom_config[param] = value end
end
end
f_in:close()
end
-- 从 smartdns 配置中读取值,优先级以 custom.conf 为准
for _, opt in ipairs(options) do
local val = custom_config[opt.config_key] or uci:get("smartdns", "@smartdns[0]", opt.key) or opt.default
if val == "yes" then val = "1" elseif val == "no" then val = "0" end
if opt.yes_no then
local arg = (val == "1" and opt.arg_yes or opt.arg_no)
if arg and arg ~= "" then
LOCAL_EXTEND_ARG = LOCAL_EXTEND_ARG .. (LOCAL_EXTEND_ARG ~= "" and " " or "") .. arg
end
else
if val and (not opt.value or (opt.invert and val ~= opt.value) or (not opt.invert and val == opt.value)) then
LOCAL_EXTEND_ARG = LOCAL_EXTEND_ARG .. (LOCAL_EXTEND_ARG ~= "" and " " or "") .. (opt.prefix or "") .. (opt.arg or val)
end
end
end
end
if not REMOTE_GROUP or REMOTE_GROUP == "nil" then
REMOTE_GROUP = "passwall_proxy"
if REMOTE_DNS then
REMOTE_DNS = REMOTE_DNS:gsub("#", ":")
end
sys.call('sed -i "/passwall/d" /etc/smartdns/custom.conf >/dev/null 2>&1')
end
local force_https_soa = uci:get(appname, "@global[0]", "force_https_soa") or 1
local proxy_server_name = "passwall-proxy-server"
config_lines = {
tonumber(force_https_soa) == 1 and "force-qtype-SOA 65" or "force-qtype-SOA -,65",
"server 114.114.114.114 -bootstrap-dns",
DNS_MODE == "socks" and string.format("proxy-server socks5://%s -name %s", REMOTE_PROXY_SERVER, proxy_server_name) or nil
}
if DNS_MODE == "socks" then
for w in string.gmatch(REMOTE_DNS, '[^|]+') do
local server_dns = api.trim(w)
local server_param
local dnsType = string.match(server_dns, "^(.-)://")
dnsType = dnsType and string.lower(dnsType) or nil
local dnsServer = string.match(server_dns, "://(.+)") or server_dns
if dnsType and dnsType ~= "" and dnsType ~= "udp" then
if dnsType == "tcp" then
server_param = "server-tcp " .. dnsServer
elseif dnsType == "tls" then
server_param = "server-tls " .. dnsServer
elseif dnsType == "quic" then
server_param = "server-quic " .. dnsServer
elseif dnsType == "https" or dnsType == "h3" then
local http_host = nil
local url = w
local port = 443
local s = api.split(w, ",")
if s and #s > 1 then
url = s[1]
local dns_ip = s[2]
local host_port = api.get_domain_from_url(s[1])
if host_port and #host_port > 0 then
http_host = host_port
local s2 = api.split(host_port, ":")
if s2 and #s2 > 1 then
http_host = s2[1]
port = s2[2]
end
url = url:gsub(http_host, dns_ip)
end
end
server_dns = url
if http_host then
server_dns = server_dns .. " -http-host " .. http_host
end
server_param = (dnsType == "https" and "server-https " or "server-h3 ") .. server_dns
end
else
server_param = "server " .. dnsServer
end
if not api.is_local_ip(w) then
server_param = server_param .. " -proxy " .. proxy_server_name
end
server_param = server_param .. " -group " .. REMOTE_GROUP .. " -exclude-default-group"
if SUBNET and SUBNET ~= "" and SUBNET ~= "0" then
server_param = server_param .. " -subnet " .. SUBNET
end
table.insert(config_lines, server_param)
end
REMOTE_FAKEDNS = 0
else
local server_param = string.format("server %s -group %s -exclude-default-group", TUN_DNS:gsub("#", ":"), REMOTE_GROUP)
table.insert(config_lines, server_param)
log(" - " .. DNS_MODE:gsub("^%l",string.upper) .. " " .. TUN_DNS .. " -> " .. REMOTE_GROUP)
end
--设置默认 DNS 分组(托底组)
local DEFAULT_DNS_GROUP = (USE_DEFAULT_DNS == "direct" and LOCAL_GROUP) or
(USE_DEFAULT_DNS == "remote" and REMOTE_GROUP)
local only_global = (DEFAULT_PROXY_MODE == "proxy" and CHN_LIST == "0" and USE_GFW_LIST == "0") and 1 --没有启用中国列表和GFW列表时(全局)
if only_global == 1 then
DEFAULT_DNS_GROUP = REMOTE_GROUP
end
if DEFAULT_DNS_GROUP then
local domain_rules_str = "domain-rules /./ -nameserver " .. DEFAULT_DNS_GROUP
if DEFAULT_DNS_GROUP == REMOTE_GROUP then
domain_rules_str = domain_rules_str .. " -speed-check-mode none -d no -no-serve-expired"
if NO_PROXY_IPV6 == "1" then
domain_rules_str = domain_rules_str .. " -address #6"
end
elseif DEFAULT_DNS_GROUP == LOCAL_GROUP then
domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "")
end
table.insert(config_lines, domain_rules_str)
end
local setflag = (NFTFLAG == "1") and "inet#passwall#" or ""
local set_type = (NFTFLAG == "1") and "-nftset" or "-ipset"
--预设排序标签(越往后优先级越高)
for i = 1, 8 do
table.insert(config_lines, "#--" .. i)
end
--屏蔽列表
local file_block_host = TMP_ACL_PATH .. "/block_host"
if USE_BLOCK_LIST == "1" and not fs.access(file_block_host) then
local block_domain, lookup_block_domain = {}, {}
local geosite_arg = ""
for line in io.lines(RULES_PATH .. "/block_host") do
if not line:find("#") and line:find("geosite:") then
line = string.match(line, ":([^:]+)$")
geosite_arg = geosite_arg .. (geosite_arg ~= "" and "," or "") .. line
else
line = api.get_std_domain(line)
if line ~= "" and not line:find("#") then
insert_unique(block_domain, line, lookup_block_domain)
end
end
end
if #block_domain > 0 then
local f_out = io.open(file_block_host, "w")
for i = 1, #block_domain do
f_out:write(block_domain[i] .. "\n")
end
f_out:close()
end
if USE_GEOVIEW == "1" and geosite_arg ~= "" and api.is_finded("geoview") then
if get_geosite(geosite_arg, file_block_host) == 0 then
log(" * 解析[屏蔽列表] Geosite 到屏蔽域名表(blocklist)完成")
else
log(" * 解析[屏蔽列表] Geosite 到屏蔽域名表(blocklist)失败!")
end
end
end
if USE_BLOCK_LIST == "1" and is_file_nonzero(file_block_host) then
local domain_set_name = "passwall-block"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, file_block_host),
string.format("domain-rules /domain-set:%s/ -a #", domain_set_name)
}
insert_array_after(config_lines, tmp_lines, "#--7")
end
--始终用国内DNS解析节点域名
local file_vpslist = TMP_ACL_PATH .. "/vpslist"
if not is_file_nonzero(file_vpslist) then
local f_out = io.open(file_vpslist, "w")
local written_domains = {}
uci:foreach(appname, "nodes", function(t)
local function process_address(address)
if address == "engage.cloudflareclient.com" then return end
if datatypes.hostname(address) and not written_domains[address] then
f_out:write(address .. "\n")
written_domains[address] = true
end
end
process_address(t.address)
process_address(t.download_address)
end)
f_out:close()
end
if is_file_nonzero(file_vpslist) then
local domain_set_name = "passwall-vpslist"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, file_vpslist)
}
local sets = {
"#4:" .. setflag .. "passwall_vps",
"#6:" .. setflag .. "passwall_vps6"
}
local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "")
domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")
domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "")
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--8")
log(string.format(" - 节点列表中的域名(vpslist)使用分组:%s", LOCAL_GROUP or "默认"))
end
--直连(白名单)列表
local file_direct_host = TMP_ACL_PATH .. "/direct_host"
if USE_DIRECT_LIST == "1" and not fs.access(file_direct_host) then
local direct_domain, lookup_direct_domain = {}, {}
local geosite_arg = ""
for line in io.lines(RULES_PATH .. "/direct_host") do
if not line:find("#") and line:find("geosite:") then
line = string.match(line, ":([^:]+)$")
geosite_arg = geosite_arg .. (geosite_arg ~= "" and "," or "") .. line
else
line = api.get_std_domain(line)
if line ~= "" and not line:find("#") then
insert_unique(direct_domain, line, lookup_direct_domain)
end
end
end
if #direct_domain > 0 then
local f_out = io.open(file_direct_host, "w")
for i = 1, #direct_domain do
f_out:write(direct_domain[i] .. "\n")
end
f_out:close()
end
if USE_GEOVIEW == "1" and geosite_arg ~= "" and api.is_finded("geoview") then
if get_geosite(geosite_arg, file_direct_host) == 0 then
log(" * 解析[直连列表] Geosite 到域名白名单(whitelist)完成")
else
log(" * 解析[直连列表] Geosite 到域名白名单(whitelist)失败!")
end
end
end
if USE_DIRECT_LIST == "1" and is_file_nonzero(file_direct_host) then
local domain_set_name = "passwall-directlist"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, file_direct_host)
}
local sets = {
"#4:" .. setflag .. "passwall_white",
"#6:" .. setflag .. "passwall_white6"
}
local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "")
domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")
domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "")
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--6")
log(string.format(" - 域名白名单(whitelist)使用分组:%s", LOCAL_GROUP or "默认"))
end
--代理(黑名单)列表
local file_proxy_host = TMP_ACL_PATH .. "/proxy_host"
if USE_PROXY_LIST == "1" and not fs.access(file_proxy_host) then
local proxy_domain, lookup_proxy_domain = {}, {}
local geosite_arg = ""
for line in io.lines(RULES_PATH .. "/proxy_host") do
if not line:find("#") and line:find("geosite:") then
line = string.match(line, ":([^:]+)$")
geosite_arg = geosite_arg .. (geosite_arg ~= "" and "," or "") .. line
else
line = api.get_std_domain(line)
if line ~= "" and not line:find("#") then
insert_unique(proxy_domain, line, lookup_proxy_domain)
end
end
end
if #proxy_domain > 0 then
local f_out = io.open(file_proxy_host, "w")
for i = 1, #proxy_domain do
f_out:write(proxy_domain[i] .. "\n")
end
f_out:close()
end
if USE_GEOVIEW == "1" and geosite_arg ~= "" and api.is_finded("geoview") then
if get_geosite(geosite_arg, file_proxy_host) == 0 then
log(" * 解析[代理列表] Geosite 到代理域名表(blacklist)完成")
else
log(" * 解析[代理列表] Geosite 到代理域名表(blacklist)失败!")
end
end
end
if USE_PROXY_LIST == "1" and is_file_nonzero(file_proxy_host) then
local domain_set_name = "passwall-proxylist"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, file_proxy_host)
}
local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP)
domain_rules_str = domain_rules_str .. " -speed-check-mode none"
domain_rules_str = domain_rules_str .. " -no-serve-expired"
local sets = {
"#4:" .. setflag .. "passwall_black"
}
if NO_PROXY_IPV6 == "1" then
domain_rules_str = domain_rules_str .. " -address #6"
domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str
else
table.insert(sets, "#6:" .. setflag .. "passwall_black6")
domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str
end
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--5")
log(string.format(" - 代理域名表(blacklist)使用分组:%s", REMOTE_GROUP or "默认"))
end
--GFW列表
if USE_GFW_LIST == "1" and is_file_nonzero(RULES_PATH .. "/gfwlist") then
local domain_set_name = "passwall-gfwlist"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, RULES_PATH .. "/gfwlist")
}
local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP)
domain_rules_str = domain_rules_str .. " -speed-check-mode none"
domain_rules_str = domain_rules_str .. " -no-serve-expired"
local sets = {
"#4:" .. setflag .. "passwall_gfw"
}
if NO_PROXY_IPV6 == "1" then
domain_rules_str = domain_rules_str .. " -address #6"
domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str
else
table.insert(sets, "#6:" .. setflag .. "passwall_gfw6")
domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str
end
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--1")
log(string.format(" - 防火墙域名表(gfwlist)使用分组:%s", REMOTE_GROUP or "默认"))
end
--中国列表
if CHN_LIST ~= "0" and is_file_nonzero(RULES_PATH .. "/chnlist") then
local domain_set_name = "passwall-chnlist"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, RULES_PATH .. "/chnlist")
}
if CHN_LIST == "direct" then
local sets = {
"#4:" .. setflag .. "passwall_chn",
"#6:" .. setflag .. "passwall_chn6"
}
local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "")
domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")
domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "")
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--2")
log(string.format(" - 中国域名表(chnroute)使用分组:%s", LOCAL_GROUP or "默认"))
end
--回中国模式
if CHN_LIST == "proxy" then
local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP)
domain_rules_str = domain_rules_str .. " -speed-check-mode none"
domain_rules_str = domain_rules_str .. " -no-serve-expired"
local sets = {
"#4:" .. setflag .. "passwall_chn"
}
if NO_PROXY_IPV6 == "1" then
domain_rules_str = domain_rules_str .. " -address #6"
domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str
else
table.insert(sets, "#6:" .. setflag .. "passwall_chn6")
domain_rules_str = REMOTE_FAKEDNS ~= "1" and (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ",")) or domain_rules_str
end
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--2")
log(string.format(" - 中国域名表(chnroute)使用分组:%s", REMOTE_GROUP or "默认"))
end
end
--分流规则
if uci:get(appname, TCP_NODE, "protocol") == "_shunt" then
local white_domain, lookup_white_domain = {}, {}
local shunt_domain, lookup_shunt_domain = {}, {}
local file_white_host = FLAG_PATH .. "/shunt_direct_host"
local file_shunt_host = FLAG_PATH .. "/shunt_proxy_host"
local geosite_white_arg, geosite_shunt_arg = "", ""
local t = uci:get_all(appname, TCP_NODE)
local default_node_id = t["default_node"] or "_direct"
uci:foreach(appname, "shunt_rules", function(s)
local _node_id = t[s[".name"]]
if _node_id and _node_id ~= "_blackhole" then
if _node_id == "_default" then
_node_id = default_node_id
end
local domain_list = s.domain_list or ""
for line in string.gmatch(domain_list, "[^\r\n]+") do
if line ~= "" and not line:find("#") and not line:find("regexp:") and not line:find("ext:") then
if line:find("geosite:") then
line = string.match(line, ":([^:]+)$")
if _node_id == "_direct" then
geosite_white_arg = geosite_white_arg .. (geosite_white_arg ~= "" and "," or "") .. line
else
geosite_shunt_arg = geosite_shunt_arg .. (geosite_shunt_arg ~= "" and "," or "") .. line
end
else
if line:find("domain:") or line:find("full:") then
line = string.match(line, ":([^:]+)$")
end
line = api.get_std_domain(line)
if line ~= "" and not line:find("#") then
if _node_id == "_direct" then
insert_unique(white_domain, line, lookup_white_domain)
else
insert_unique(shunt_domain, line, lookup_shunt_domain)
end
end
end
end
end
if _node_id ~= "_direct" then
log(string.format(" - Sing-Box/Xray分流规则(%s)使用分组:%s", s.remarks, REMOTE_GROUP or "默认"))
end
end
end)
if is_file_nonzero(file_white_host) == nil then
if #white_domain > 0 then
local f_out = io.open(file_white_host, "w")
for i = 1, #white_domain do
f_out:write(white_domain[i] .. "\n")
end
f_out:close()
end
end
if is_file_nonzero(file_shunt_host) == nil then
if #shunt_domain > 0 then
local f_out = io.open(file_shunt_host, "w")
for i = 1, #shunt_domain do
f_out:write(shunt_domain[i] .. "\n")
end
f_out:close()
end
end
if USE_GFW_LIST == "1" and CHN_LIST == "0" and USE_GEOVIEW == "1" and api.is_finded("geoview") then --仅GFW模式解析geosite
local return_white, return_shunt
if geosite_white_arg ~= "" then
return_white = get_geosite(geosite_white_arg, file_white_host)
end
if geosite_shunt_arg ~= "" then
return_shunt = get_geosite(geosite_shunt_arg, file_shunt_host)
end
if (return_white == nil or return_white == 0) and (return_shunt == nil or return_shunt == 0) then
log(" * 解析[分流节点] Geosite 完成")
else
log(" * 解析[分流节点] Geosite 失败!")
end
end
if is_file_nonzero(file_white_host) then
local domain_set_name = "passwall-whitehost"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, file_white_host)
}
local domain_rules_str = string.format('domain-rules /domain-set:%s/ %s', domain_set_name, LOCAL_GROUP and "-nameserver " .. LOCAL_GROUP or "")
if USE_DIRECT_LIST == "1" then
local sets = {
"#4:" .. setflag .. "passwall_white",
"#6:" .. setflag .. "passwall_white6"
}
domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")
else
local sets = {
"#4:" .. setflag .. "passwall_shunt",
"#6:" .. setflag .. "passwall_shunt6"
}
domain_rules_str = domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ",")
end
domain_rules_str = domain_rules_str .. (LOCAL_EXTEND_ARG ~= "" and " " .. LOCAL_EXTEND_ARG or "")
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--4")
end
if is_file_nonzero(file_shunt_host) then
local domain_set_name = "passwall-shuntlist"
tmp_lines = {
string.format("domain-set -name %s -file %s", domain_set_name, file_shunt_host)
}
local domain_rules_str = string.format('domain-rules /domain-set:%s/ -nameserver %s', domain_set_name, REMOTE_GROUP)
domain_rules_str = domain_rules_str .. " -speed-check-mode none"
domain_rules_str = domain_rules_str .. " -no-serve-expired"
local sets = {
"#4:" .. setflag .. "passwall_shunt"
}
if NO_PROXY_IPV6 == "1" then
domain_rules_str = domain_rules_str .. " -address #6"
domain_rules_str = (not only_global and REMOTE_FAKEDNS == "1")
and domain_rules_str
or (domain_rules_str .. " " .. set_type .. " " .. table.concat(sets, ","))
else
table.insert(sets, "#6:" .. setflag .. "passwall_shunt6")
domain_rules_str = (not only_global and REMOTE_FAKEDNS == "1")
and domain_rules_str
or (domain_rules_str .. " -d no " .. set_type .. " " .. table.concat(sets, ","))
end
table.insert(tmp_lines, domain_rules_str)
insert_array_after(config_lines, tmp_lines, "#--3")
end
end
if #config_lines > 0 then
local f_out = io.open(TMP_CONF_FILE, "w")
for i = 1, #config_lines do
line = config_lines[i]
if line ~= "" and not line:find("^#--") then
f_out:write(line .. "\n")
end
end
f_out:close()
end
if DEFAULT_DNS_GROUP then
log(string.format(" - 默认 DNS 分组:%s", DEFAULT_DNS_GROUP))
end
fs.symlink(TMP_CONF_FILE, SMARTDNS_CONF)
sys.call(string.format('echo "conf-file %s" >> /etc/smartdns/custom.conf', string.gsub(SMARTDNS_CONF, appname, appname .. "*")))
log(" - 请让SmartDNS作为Dnsmasq的上游或重定向!")
|
294coder/Efficient-MIF | 34,158 | utils/log_utils.py | import os
import json
import logging
from datetime import datetime
from functools import partial
import signal
from typing import Any, Dict, List, Optional, Union, Protocol
from collections.abc import Sequence, Iterable
from contextlib import contextmanager
import loguru
import matplotlib.pyplot as plt
import numpy as np
import torch
from torch.utils.tensorboard import SummaryWriter
from torch import nn
from torchvision.utils import make_grid
import shortuuid
from utils.misc import NameSpace, default, is_main_process
from utils.visualize import get_spectral_image_ready
import time
from rich.console import Console
from rich.logging import RichHandler
from importlib import reload
def get_time(sec):
h = int(sec//3600)
m = int((sec//60)%60)
s = int(sec%60)
return h,m,s
class TimeFilter(logging.Filter):
def filter(self, record):
try:
start = self.start
except AttributeError:
start = self.start = time.time()
time_elapsed = get_time(time.time() - start)
record.relative = "{0}:{1:02d}:{2:02d}".format(*time_elapsed)
# self.last = record.relativeCreated/1000.0
return True
def save2json_file(d: dict, path: str, mode: str = "w", indent: int = 4):
assert path.endswith(".json"), "@path should end with .json"
with open(path, mode) as f:
json.dump(d, f, indent=indent)
print(f"save json file in {path}")
def loss_dict2str(
loss_dict: "dict[int, float | torch.Tensor]",
world_size: int = 1,
round_fp: int = 6
) -> str:
log_str = ""
for k, v in loss_dict.items():
log_str += f"{k}: {v / world_size:.{round_fp}f} "
return log_str
## decrepted
class TrainStatusLogger(object):
def __init__(
self, id="None", path="./train_status/status.pt", resume=False, args=None
):
"""
track training status as a context manager
:param id: run id, which is defined in parser args
:param path: pkl file's path
:param resume: resume training. if you want to resume last training run, set id=None.
if you want to specify one run to resume, pass a specified id.
"""
self.time_now = datetime.now()
self.path = path
self.id = id
self.resume = resume
self.status = {
"id": id,
"status": "running",
"time_stamp": str(self.time_now.timestamp()),
"time": self.time_now.strftime("%Y-%m-%d, %H:%M:%S"),
"args": args,
}
self._base_path = os.path.dirname(self.path)
if not os.path.exists(self.path):
if not os.path.exists(self._base_path):
os.mkdir(self._base_path)
self.status_all = [self.status]
else:
self.status_all = self.load_train_status()
self.check_unique_id()
if resume:
if id == "None":
self.status, _ = self.find_last_untrained_status()
else:
self.status = self.find_run_by_id(id)
# print('warning: argument @id is not equal to the resume id which will be ignored')
else:
self.status_all.append(self.status)
# ======= handle the KeyboardInterrupt signal =======
def handler(*args):
print('catch signal: KeyboardInterrupt')
print('EXITTING...')
raise KeyboardInterrupt
signal.signal(signal.SIGINT, handler)
@staticmethod
def _check_status_legal(status):
assert status in ("running", "done", "break")
def load_train_status(self):
if os.path.getsize(self.path) > 0:
# with open(self.path, "rb") as f:
# l = pickle.load(f)
l = torch.load(self.path)
else:
raise EOFError("file is empty, you should delete it")
print("load previous train status")
return l
def save_train_status(self):
# with open(self.path, "wb") as f:
# pickle.dump(self.status_all, f)
torch.save(self.status_all, self.path)
print("save all train status")
def update_train_status(self, status):
self._check_status_legal(status)
self.status["status"] = status
def find_last_untrained_status(self):
f_sort = lambda d: d["time_stamp"] if d["status"] == "break" else "0"
last_status = sorted(self.status_all, key=f_sort)[-1]
return last_status, last_status["id"]
def find_run_by_id(self, id):
s = self._find_id(id)
if s["status"] != "break":
return s
raise AttributeError(
f"no id: {id} in not an existing run or has already been done"
)
def _find_id(self, id):
for s in self.status_all:
if s["id"] == id:
return s
def print_status_by_id(self, id):
s = self._find_id(id)
for k, v in s:
if isinstance(v, NameSpace):
print(v)
else:
print(f"{k}: {v}")
def check_unique_id(self):
ids = []
for d in self.status_all:
ids.append(d["id"])
assert len(ids) == len(np.unique(ids)), "exist id conflict"
assert not self.status["id"] in ids or self.resume, (
"id conflicts, check your run id "
"or delete all tracker pkl file. "
f"the pkl file can be found in {self.path}"
)
def __enter__(self):
nbreak = 0
ndone = 0
for d in self.status_all:
s = d["status"]
if s == "done":
ndone += 1
elif s == "break":
nbreak += 1
print("=" * 20, "Log Train Process", "=" * 20, sep="")
print(f"all runs: {ndone} run(s) done, {nbreak} run(s) break")
return self
def __exit__(self, exc_type, exc_val, exc_tb):
# print(f'traceback: {exc_tb}')
if exc_type is not None or exc_val is not None:
print("=" * 20, "Find Error Happen", "=" * 20, sep="")
print(f"catch error type: {exc_type}, error value: {exc_val}")
self.update_train_status("break")
else:
print("=" * 20, "Training End", "=" * 20, sep="")
self.update_train_status("done")
# only save in main process
if is_main_process():
self.save_train_status()
def __repr__(self):
def dict_str(d):
s = "id: {:<10} status: {:<7} time_stamp: {:<20} time: {:<20}".format(
d["id"], d["status"], d["time_stamp"], d["time"]
)
return s
repr = ""
for d in self.status_all:
repr += dict_str(d) + "\n"
return repr
def generate_id(length: int = 8) -> str:
# ~3t run ids (36**8)
run_gen = shortuuid.ShortUUID(alphabet=list("0123456789abcdefghijklmnopqrstuvwxyz"))
return str(run_gen.random(length))
def easy_logger(level='INFO', format_str: str=None, func_name: str=None):
reload(logging)
if format_str is not None and func_name is not None:
format_str = "%(func_name)s: " + format_str
elif format_str is None and func_name is not None:
format_str = "%(func_name)s: %(message)s"
else:
format_str = "%(message)s"
file_name = os.environ.get('LOG_FILE', None)
if file_name:
console = Console(file=open(file_name, 'a+'))
else:
console = None
class FuncNameFilter(logging.Filter):
def filter(self, record):
record.func_name = func_name
return True
rich_handler = RichHandler(console=console, show_path=False, level=level, markup=True)
rich_handler.setFormatter(logging.Formatter(format_str, datefmt='%X'))
logger = logging.getLogger(__name__)
logger.setLevel(level)
logger.addHandler(rich_handler)
class ProtocalLogger(Protocol):
@classmethod
def print(ctx, *msg, sep=" ", level: str | int="INFO"):
msg = sep.join(map(str, msg))
if isinstance(level, str):
level = eval(f'logging.{level.upper()}')
logger.log(level, msg, extra={"markup": True})
@classmethod
def debug(ctx, msg):
pass
@classmethod
def info(ctx, msg):
pass
@classmethod
def warning(ctx, msg):
pass
@classmethod
def error(ctx, *msg, sep=" ", raise_error: bool=False, error_type=None):
msg = sep.join(map(str, msg))
ctx.print(msg, level='ERROR')
if raise_error:
if error_type is not None:
raise error_type(msg)
raise RuntimeError(msg)
# add filter to add func_name to log
if func_name is not None:
logger.addFilter(FuncNameFilter())
# logger signature
logger: ProtocalLogger
# add attributes to logger
logger.print = ProtocalLogger.print
logger.debug = partial(ProtocalLogger.print, level='DEBUG')
logger.info = partial(ProtocalLogger.print, level='INFO')
logger.warning = partial(ProtocalLogger.print, level='WARNING')
logger.error = ProtocalLogger.error
logger._console = rich_handler.console
return logger
from loguru import logger, _logger
class LoguruLogger:
_logger = logger
console = None
handler = []
_first_import = True
_default_file_format = "<green>[{time:MM-DD HH:mm:ss}]</green> <level>[{level}] {message}</level>"
_default_console_format = "[{time:HH:mm:ss}] <level>[{level}] {message}</level>"
@classmethod
def logger(cls,
sink=None,
format=None,
filter=None,
**kwargs) -> "_logger.Logger":
reload(loguru)
if cls._first_import:
cls._logger.remove() # the first time import
cls.console = Console(color_system=None)
cls._logger.add(
default(sink, lambda x: cls.console.print(x)),
colorize=True,
format=default(format, cls._default_console_format),
**kwargs
)
cls._first_import = False
else:
if sink is not None:
handler = cls._logger.add(sink, format=default(format, cls._default_file_format), filter=filter, **kwargs)
cls.handler.append(handler)
return cls._logger
@classmethod
def add(cls, *args, **kwargs):
handler = cls._logger.add(*args, **kwargs)
cls.handler.append(handler)
@classmethod
def remove_all(cls):
for h in cls.handler:
cls._logger.remove(h)
cls.handler = []
@classmethod
def remove_id(cls, id):
cls._logger.remove(id)
@classmethod
def bind(cls, *args, **kwargs):
return cls._logger.bind(*args, **kwargs)
@contextmanager
def catch_any_error():
try:
logger = LoguruLogger.logger()
yield logger
except Exception as e:
logger.error(f"catch error: {e}", raise_error=True)
logger.exception(e)
finally:
LoguruLogger.remove_all()
def get_logger(
base_path: str = None,
name: str = None,
args=None,
std_level=logging.INFO,
file_level: Union[tuple, int] = (logging.DEBUG,),
file_handler_names: Union[tuple, str] = ("debug",),
file_mode: str = "w",
show_pid: bool = False,
method_dataset_as_prepos=True,
):
"""
get logger to log
:param base_path: such like './log/'
:param name: logger name such as 'train_epoch_300'
:param std_level: stream level
:param file_level: file level
:param file_handler_names:
:param file_mode: 'a' append, 'w' write
:param show_pid: show thread id
:return: logger and List[handlers]
"""
assert name is not None, "@param name should not be None"
reload(logging)
# assert base_path is not None, "@param base_path should not be None"
if not show_pid:
# format_str = "[%(asctime)s - %(funcName)s]-%(levelname)s: %(message)s"
# format_str = "(%(relative)s) %(message)s"
format_str = "%(message)s"
else:
# format_str = (
# "[%(asctime)s - %(funcName)s - pid: %(thread)d]-%(levelname)s: %(message)s"
# )
format_str = "pid: %(thread)d) %(message)s"
rich_handler = RichHandler(show_path=False, markup=True)
rich_handler.setFormatter(logging.Formatter(format_str, datefmt='%X'))
logger = logging.getLogger(name)
logger.setLevel(logging.DEBUG)
logger.addHandler(rich_handler)
hdls = []
# stream_handler = logging.StreamHandler(sys.stdout)
# stream_handler.setLevel(std_level)
# hdls.append(stream_handler)
# logger.addHandler(stream_handler)
if base_path is not None:
assert len(file_handler_names) == len(file_level), \
"@param file_handler_names and @param file_level should be list and equal length"
for n, level in zip(file_handler_names, file_level):
if method_dataset_as_prepos:
file_log_dir = os.path.join(base_path, args.full_arch, args.dataset, name)
else:
file_log_dir = os.path.join(base_path, name)
if not os.path.exists(file_log_dir):
os.makedirs(file_log_dir)
print(f"logging: make log file [{os.path.abspath(file_log_dir)}]")
file_log_path = os.path.join(file_log_dir, n + ".log")
# file_handler = logging.FileHandler(file_log_path, mode=file_mode)
file_console = Console(file=open(file_log_path, 'w'))
file_handler = RichHandler(console=file_console, show_path=False, markup=True)
# formatter = logging.Formatter(
# "[%(asctime)s - %(name)s] - %(levelname)s: %(message)s"
# )
# file_handler.setFormatter(formatter)
file_handler.setLevel(level)
hdls.append(file_handler)
logger.addHandler(file_handler)
else:
file_console = None
file_handler = None
file_log_dir = None
file_log_path = None
for handler in logger.handlers:
handler.addFilter(TimeFilter())
def log_print(*msg, sep=" ", level="INFO", dist=False, proc_id=None):
if dist or is_main_process():
if isinstance(level, str):
level_int = eval(f"logging.{level}")
msgs = f"{proc_id=} - " if proc_id is not None else ""
msgs += sep.join(map(str, msg))
logger.log(level=level_int, msg=msgs)
# register some attrs
logger.print = log_print
logger._console = rich_handler.console
logger._file_console = file_console
logger._file_path = file_log_path
return logger, hdls, file_log_dir
# class WandbLogger:
# def __init__(
# self,
# project_name,
# run_name=None,
# save_path=None,
# config=None,
# resume=None,
# **kwargs,
# ):
# self.run = wandb.init(
# project=project_name,
# name=run_name,
# dir=save_path,
# config=config,
# resume=resume,
# settings=wandb.Settings(start_method="fork"),
# **kwargs,
# )
# def watch(self, type, network, graph=False, freq=20):
# # type(str) One of "gradients", "parameters", "all", or None
# wandb.watch(network, log=type, log_graph=graph, log_freq=freq)
# def log_image(self, tensor_or_numpy, img_name, *args):
# # e.g. cols = ['id','pred', 'gt', 'res']
# # tensor_or_numpy should be tensor [B, C, H, W] or numpy array [H, W, C]
# # C can only be 1, 3 or 4
# # warning: you should control B for clear visualization, suggest 2<B<7
# if isinstance(tensor_or_numpy, torch.Tensor):
# x = tensor_or_numpy.cpu().numpy()
# elif isinstance(tensor_or_numpy, np.ndarray):
# x = tensor_or_numpy
# else:
# raise NotImplementedError
# assert x.shape[-1] in [1, 3, 4]
# img_log = [wandb.Image(xi) for xi in x] if x.ndim == 4 else [wandb.Image(x)]
# self.run.log({img_name: img_log})
# def log_table(self, table_data, cols, table_name, *args):
# # for example
# # my_data = [
# # [0, wandb.Image("img_0.jpg"), 0, 0],
# # [1, wandb.Image("img_1.jpg"), 8, 0],
# # [2, wandb.Image("img_2.jpg"), 7, 1],
# # [3, wandb.Image("img_3.jpg"), 1, 1]
# # ]
# #
# # create a wandb.Table() with corresponding columns
# # columns =["id", "image", "prediction", "truth"]
# # test_table = wandb.Table(data=my_data, columns=columns)
# # add_data
# # add row: table.add_data("3a", "3b", "3c")
# # add col: table.add_column(name="col_name", data=col_data)
# wandb.log({table_name: wandb.Table(data=table_data, columns=cols)})
# def log_curve(self, value=None, name=None, d=None, *args):
# if d is None:
# wandb.log({name: value})
# else:
# wandb.log(d)
class NoneLogger:
def __init__(self, *args, **kwargs):
class NoneWriter:
def __init__(self) -> None:
pass
def close(self, *args, **kwargs):
pass
self.writer = NoneWriter()
self.logger, *_ = get_logger(name=kwargs.pop('name'))
# add time and run_id
args = kwargs['cfg']
args.logger_config.name = (
time.strftime("%Y-%m-%d-%H-%M-%S", time.localtime())
+ "_"
+ args.logger_config.name
)
args.logger_config.name += "_" + args.run_id + f"_{args.comment}"
name = args.logger_config.name
self.log_file_dir = os.path.join(args.logger_config.base_path,
args.full_arch, args.dataset, name)
@property
def console(self):
return self.logger._console
@property
def file_console(self):
return self.file_console._file_console
def watch(self, *args, **kwargs):
pass
def log_image(self, *args, **kwargs):
pass
def log_images(self, *args, **kwargs):
pass
def log_curve(self, *args, **kwargs):
pass
def log_curves(self, *args, **kwargs):
pass
def log_network(self, *args, **kwargs):
pass
def print(self, *msg, level="INFO", dist=False, proc_id=None):
if dist or is_main_process():
if isinstance(level, str):
level_int = eval(f"logging.{level}")
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.logger.log(level=level_int, msg=msgs)
def info(self, *msg, dist=False, proc_id=None):
if dist or is_main_process():
level_int = logging.INFO
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.logger.log(level=level_int, msg=msgs)
def debug(self, *msg, dist=False, proc_id=None):
if dist or is_main_process():
level_int = logging.DEBUG
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.logger.log(level=level_int, msg=msgs)
def warning(self, *msg, dist=False, proc_id=None):
if dist or is_main_process():
level_int = logging.WARNING
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.logger.log(level=level_int, msg=msgs)
def error(self, *msg, dist=False, proc_id=None):
if dist or is_main_process():
level_int = logging.ERROR
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.logger.log(level=level_int, msg=msgs)
class TensorboardLogger:
def __init__(
self,
args=None,
tsb_logdir=None,
comment=None,
file_stream_log=True,
config_file_mv="./configs",
config_file_type="yaml",
method_dataset_as_prepos=False
):
"""
Args:
args: config args from main.py
tsb_logdir: tensorboard log dir
comment:
file_stream_log: file stream dir
config_file_mv: where arch_config.yaml dir at
"""
self.grad_dict = {}
# if not os.path.exists(logdir):
# os.mkdir(logdir)
self.hooks = {}
self.watch_type = "None"
self.freq = 10
# add time and run_id
args.logger_config.name = (
time.strftime("%Y-%m-%d-%H-%M-%S", time.localtime())
+ "_"
+ args.logger_config.name
)
args.logger_config.name += "_" + args.run_id + f"_{args.comment}"
if file_stream_log:
self.file_logger, self.file_hdls, self.log_file_dir = get_logger(
**args.logger_config.to_dict(),
args=args,
method_dataset_as_prepos=method_dataset_as_prepos
)
config_cp_path = os.path.join(self.log_file_dir, "config.json")
save2json_file(args.to_dict(), config_cp_path)
# shutil.copy2(os.path.join(config_file_mv, f'{args.arch}_config.{config_file_type}'), self.log_file_dir)
self.print(
f"\nmove config file to {os.path.abspath(self.log_file_dir)}", level="INFO"
)
self.writer = SummaryWriter(default(tsb_logdir, self.log_file_dir), comment)
# set file path to os.environ
os.environ['LOG_FILE'] = self.logger_file_path
self.file_logger.debug(f'set os environ `LOG_FILE`={self.logger_file_path}')
@property
def console(self) -> Console:
return self.file_logger._console
@property
def logger_file_path(self) -> str:
return self.file_logger._file_path
@property
def file_console(self):
return self.file_console._file_console
def check_tensor_float(self, x):
if isinstance(x, torch.Tensor):
if x.dtype != torch.float32:
x = x.to(dtype=torch.float32)
return x
@is_main_process
def watch(self, network: nn.Module, watch_type: str, freq: int):
assert watch_type in (
"all",
"grad",
"None",
), "@watch_type should only be all, grad or None"
if watch_type == "None":
return
self.watch_type = watch_type
self.freq = freq
def _hook(grad, name):
self.grad_dict[name] = grad
for n, p in network.named_parameters():
hook = partial(_hook, name=n)
self.hooks[n] = hook
p.register_hook(hook)
@is_main_process
def log_curve(self, x, name, epoch):
self.writer.add_scalar(name, self.check_tensor_float(x), epoch)
@is_main_process
def log_curves(self, x_dict: Dict, epoch):
# for example:
# for i in range(100):
# writer.add_scalars('run_14h', {'xsinx': i * np.sin(i / r),
# 'xcosx': i * np.cos(i / r),
# 'tanx': np.tan(i / r)}, i)
# self.writer.add_scalars(main_name, x_dict, epoch)
for k, v in x_dict.items():
self.writer.add_scalar(k, self.check_tensor_float(v), epoch)
@is_main_process
def log_image(self, image, name, epoch):
if image.ndim == 3:
assert image.shape[0] <= 3, (
f"the number of image channel "
f"should not greater than 3 but got shape {image.shape}"
)
self.writer.add_image(name, self.check_tensor_float(image), epoch, dataformats="CHW")
@is_main_process
def log_images(self, batch_imgs: Sequence, nrow: int, names: Sequence, task: str,
epoch: int, ds_name: str, **grid_kwargs):
assert task in ['fusion', 'sharpening'], '@task should be fusion or sharpening'
for batch_img, name in zip(batch_imgs, names):
batch_img = get_spectral_image_ready(self.check_tensor_float(batch_img), name, task, ds_name)
grid_img = make_grid(batch_img, nrow=nrow, **grid_kwargs)
self.log_image(grid_img, name, epoch)
@is_main_process
def log_network(self, network: nn.Module, ep: int):
if self.watch_type != "None":
if ep % self.freq == 0:
for (_, g), (n, p) in zip(
self.grad_dict.items(), network.named_parameters()
):
if self.watch_type == "all":
self.writer.add_histogram(n + "_data", p.flatten().float(), ep)
self.writer.add_histogram(n + "_grad", g.flatten().float(), ep)
def print(self, *msg, level="INFO", dist=False, proc_id=None):
if dist or is_main_process():
if isinstance(level, str):
level_int = eval(f"logging.{level}")
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.file_logger.log(level=level_int, msg=msgs)
def info(self, *msg, dist=False, proc_id=None):
if dist or is_main_process():
level_int = logging.INFO
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.file_logger.log(level=level_int, msg=msgs)
def debug(self, *msg, dist=False, proc_id=None):
if dist or is_main_process():
level_int = logging.DEBUG
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.file_logger.log(level=level_int, msg=msgs)
def warning(self, *msg, dist=False, proc_id=None):
if dist or is_main_process():
level_int = logging.WARNING
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.file_logger.log(level=level_int, msg=msgs)
def error(self, *msg, dist=False, proc_id=None, raise_error: bool=False, error_type=None):
if dist or is_main_process():
level_int = logging.ERROR
msgs = f"{proc_id=} - " if proc_id is not None else ""
for s in msg:
msgs += s
self.file_logger.log(level=level_int, msg=msgs)
if raise_error:
if error_type is not None:
raise error_type(msgs)
raise RuntimeError(msgs)
# from aim import Distribution, Image, Repo, Run, Session, Text
# class AimLogger(object):
# def __init__(
# self, run_name, resume_hash_name=None, hparams=None, cap_term_logs=True,
# *,
# desp=None,
# ):
# """Aim framewarke logger
# Args:
# run_name (str): experiment name
# resume_hash_name (str, optional): set it if you want to resume training. Defaults to None.
# hparams (dict, optional): dict of some hyperparameters of your experiments. Defaults to None.
# cap_term_logs (bool, optional): capture terminal logs. Defaults to True.
# desp (str, optional): short description of your experiment. Defaults to None.
# """
# self.run = Run(
# resume_hash_name,
# experiment=run_name,
# capture_terminal_logs=cap_term_logs,
# log_system_params=False,
# )
# setattr(self.run, 'description', desp)
# self.run["hparams"] = hparams
# self.log_info("log params:", hparams)
# self.run_name = run_name
# self.repo = Repo("./")
# def _convert_uint8_img(self, img):
# # TODO: to suit PIL package, convert the image into unit8 type
# # ref to https://pillow.readthedocs.io/en/stable/handbook/concepts.html
# # Aim package only support uint8 type image :>
# if isinstance(img, torch.Tensor):
# img = img.detach().cpu().numpy()
# img = img - img.min()
# img = img / img.max()
# img = (img * 255).astype(np.uint8)
# return img
# def log_image(self, image, name=None, epoch=None, context=None):
# # check image type
# if isinstance(image, np.ndarray):
# assert image.ndim in [3, 2]
# img = Image(self._convert_uint8_img(image))
# # TODO: convert the Tensor into ndarray is really slow
# elif isinstance(image, torch.Tensor):
# assert image.ndim in [4, 3, 2]
# if image.ndim == 4:
# nrows = math.sqrt(image.shape[0])
# image = make_grid(image, nrow=nrows)
# img = Image(image)
# elif image.ndim in [3, 2]:
# img = Image(self._convert_uint8_img(image))
# else:
# self.run.log_warning(f"not support image shape {image.shape}")
# elif isinstance(image, plt.Figure):
# img = Image(image)
# else:
# img = image
# self.run.track(img, name=name, epoch=epoch, context=context)
# def log_text(self, text, name=None, epoch=None, context=None):
# self.run.track(Text(text), name=name, epoch=epoch, context=context)
# @beartype.beartype()
# def log_metrics(self, metrics, name=None, epoch=None, context: dict = None):
# """log metrics or other values
# Args:
# metrics (dict or values): a dict or values to log
# epoch (_type_): _description_
# context (_type_, optional): _description_. Defaults to None.
# """
# self.run.track(metrics, name=name, epoch=epoch, context=context)
# @beartype.beartype()
# def log_distribution(
# self,
# distribution,
# name=None,
# epoch=None,
# context: dict = None,
# ):
# #######################################################################
# # !!!!
# # warning: the context must be a dict or it will explode your aim repo
# # I don't know why, maybe it is a bug
# #######################################################################
# distribution = distribution.flatten()
# if isinstance(distribution, torch.Tensor):
# distribution = distribution.detach().cpu().numpy()
# # hist, bins = np.histogram(distribution, bins=64 if 64 < len(distribution) else len(distribution))
# # bin range is
# # bin_range = [bins[0], bins[-1]]
# self.run.track(
# Distribution(distribution), name=name, epoch=epoch, context=context
# )
# def log_network(self, network: nn.Module, epoch: int = None):
# # refer to the warning in @log_distribution, it's important
# # I set the @context into a dict, do not change it
# for n, p in network.named_parameters():
# p = p.flatten().detach().cpu().numpy()
# # context = {"net_params_dist": n}
# self.log_distribution(
# p, name="network_params", epoch=epoch, context={"net_params_dist": n} # do not change
# )
# def close(self):
# # close the run
# # finilize and close, may one of them will take effect
# self.run.finalize()
# self.run.close()
# print("logger closed")
# def _make_msg_one_text(self, *msg):
# if isinstance(msg[0], str) and len(msg) == 1:
# return msg[0]
# fin_msg = ""
# for m in msg:
# fin_msg += str(m)
# return fin_msg
# ######## override those functions #########
# def log_info(self, *msg):
# self.run.log_info(self._make_msg_one_text(*msg))
# def log_warning(self, *msg):
# self.run.log_warning(self._make_msg_one_text(*msg))
# def log_error(self, *msg):
# self.run.log_error(self._make_msg_one_text(*msg))
# def log_debug(self, *msg):
# self.run.log_debug(self._make_msg_one_text(*msg))
# ###########################################
# # repo control
# # may be not used
# def delete_run(self, run_hash=None):
# # sometimes the run is locked, so we need to release it
# # or just delete it
# print(
# "warning, deleting run: {}".format(run_hash if run_hash else self.run.hash)
# )
# run = self.repo.get_run(run_hash if run_hash else self.run.hash)
# # run._lock.release() # will raise Nonetype do not have the attribute
# run.read_only = False
# ans = input("press [y/n] to confirm deleting")
# if ans == "y":
# d_m = self.repo.delete_run(run_hash)
# print("deleted {}: {}".format(run.hash, d_m))
# elif ans == "n":
# print("canceled")
# else:
# print("invalid input, canceled")
# @property
# def hash(self):
# return self.run.hash
if __name__ == '__main__':
logger = easy_logger()
logger.info('[dim green] 1234 [/dim green]')
|
294coder/Efficient-MIF | 10,992 | utils/visualize.py | import os
import os.path as osp
from typing import Literal, Tuple, Union
import cv2
import numpy as np
import torch
from torch import Tensor
import matplotlib.pyplot as plt
from utils.misc import to_numpy
def get_rgb_channel_by_dataset_name(tensor, dataset_name: str):
if dataset_name in ('wv3', 'wv2'):
return tensor[:, [4,2,0], ...]
elif dataset_name in ('gf2', 'qb'):
return tensor[:, :3, ...]
elif dataset_name in ('gf5', 'gf5-gf1'):
return tensor[:, [40, 30, 20], ...]
elif dataset_name == 'houston':
return tensor[:, [39, 29, 19], ...]
elif 'cave' in dataset_name or 'harvard' in dataset_name:
return tensor[:, [29, 19, 9], ...]
else:
return tensor[:, :3, ...]
def permute_dim(*args):
d = [
i.permute(2, 0, 1) for i in args
]
return d
def normalize(img, to_uint8=True):
"""
centering image to show
:param img: numpy array, shape [H, W, C]
:return: uint8 type image
"""
img = img - img.min((0, 1))
img = img / img.max((0, 1))
if to_uint8:
img *= 255
img = img.astype('uint8')
return img
def invert_normalized(norm_img: Union[Tensor, np.ndarray],
mean: Union[Tensor, np.ndarray],
std: Union[Tensor, np.ndarray],
*,
change_back_dim=True):
"""
invert image normalized to unnormalized image
Args:
norm_img: Tensor: [C, H, W] or [B, C, H, W]
mean: Tensor, [C, ]
std: Tensor, [C, ]
change_back_dim: bool, change channel dim back
Returns: Ndarray, unnormalized image
"""
if isinstance(norm_img, Tensor):
norm_img = to_numpy(norm_img)[0]
if isinstance(mean, Tensor):
mean = to_numpy(mean)[0]
if isinstance(std, Tensor):
std = to_numpy(std)[0]
if norm_img.ndim == 4:
_dim_trans = [0, 2, 3, 1]
_dim_back_trans = [0, -1, 1, 2]
else:
_dim_trans = [1, 2, 0]
_dim_back_trans = [2, 0, 1]
norm_img = norm_img.transpose(_dim_trans)
assert norm_img.shape[-1] == mean.size == std.size
unnormed_img = norm_img * std + mean # [H, W, C] or [B, H, W, C]
if change_back_dim:
unnormed_img = unnormed_img.transpose(_dim_back_trans)
return unnormed_img
def hist_equal(img):
"""
equalize an image
:param img: numpy array, shape [H, W, C], C can be any int
:return:
"""
if img.ndim == 3:
for i in range(img.shape[-1]):
img[..., i] = cv2.equalizeHist(img[..., i])
else:
img = cv2.equalizeHist(img)
return img
def res_image(gt: Tensor, sr: Tensor, *, exaggerate_ratio: int = None) -> torch.Tensor:
# shape [B, C, H, W]
ratio = exaggerate_ratio if exaggerate_ratio is not None else 1.
res = torch.abs(gt - sr).mean(1, keepdim=True) * ratio
return res
def get_spectral_image_ready(batch_image: Tensor,
tensor_name: str,
task: str=None,
ds_name: Literal['wv3', 'wv2', 'gf2', 'qb',
'gf5', 'gf5-gf1', 'flir', 'tno',
'msrs']=None) -> Tensor:
# batch_image: [B, C, H, W]
img_arrs = batch_image.permute(0, 2, 3, 1).detach().cpu().numpy() # [B, H, W, C]
if task == 'fusion':
transform_fn = lambda x, const: torch.tensor(x, dtype=torch.float32)
elif task == 'sharpening':
if tensor_name in ('lms', 'ms', 'sr'): # for pansharpening and HISR tasks
batch_image = get_rgb_channel_by_dataset_name(batch_image, ds_name)
elif tensor_name == 'pan' and batch_image.shape[1] > 3:
batch_image = batch_image[:, :3]
transform_fn = lambda x, const: torch.tensor(normalize(x, to_uint8=False) * const)
else:
raise ValueError(f'Invalid task: {task}')
if 'res' in tensor_name:
equalized_img = [transform_fn(i, 10).permute(-1, 0, 1)[None, ...] for i in img_arrs] # [1, C, H, W]
else:
equalized_img = [transform_fn(i, 1).permute(-1, 0, 1)[None, ...] for i in img_arrs]
grid = torch.cat(equalized_img, dim=0)
return grid
def viz_batch(img: Tensor, base_path='./visualized_img', suffix=None, start_index=1, format='jpg'):
assert suffix is not None, 'arg @suffix can not be None'
assert suffix in ['pan', 'ms', 'sr', 'gt', 'residual'], 'arg @suffix should only be pan, ms or sr'
img_arrs = img.permute(0, 2, 3, 1).numpy()
if suffix == 'residual':
equalized_img = [i for i in img_arrs]
else:
# equalized_img = [hist_equal(normalize(i)) for i in img_arrs]
equalized_img = [normalize(i) for i in img_arrs]
path = osp.join(base_path, suffix)
if not osp.exists(path):
os.makedirs(path)
# all_path = path + '.mat'
# savemat(all_path, {f'{suffix}': img_arrs})
# fig, ax = plt.subplots(figsize=(5, 5), dpi=100)
# ax: plt.Axes
# fig: plt.Figure
for i, img in enumerate(equalized_img, start_index):
h, w = img.shape[-2:]
img_path1 = osp.join(path, str(i) + '.' + format)
# plt.cla()
if suffix == 'pan' or suffix == 'residual':
# ax.imshow(img, cmap='gray')
cv2.imwrite(img_path1, img)
# elif suffix == 'residual':
# ax.imshow(img)
else:
try:
# ax.imshow(img[..., [0, 2, 4]])
cv2.imwrite(img_path1, img[..., [0, 2, 4]])
except:
# ax.imshow(img[..., :3])
cv2.imwrite(img_path1, img[..., :3])
# ax.set_axis_off()
# fig.set_size_inches(h, w)
# fig.savefig(img_path1, format=format, dpi=50, bbox_inches='tight', pad_inches=0.)
# plt.close()
def show_details(img: np.ndarray,
cpos_ratio: Tuple[float, float],
area_pixels: Tuple[int, int],
interp_ratio: int = 3,
color: Tuple[int, int, int] = (0, 255, 0),
thickness: int = 2,
place: str = None) -> np.ndarray:
"""select a patch in raw image which decided by @cpos_ratio and @area_pixels,
the function will interpolate @interp_ratio times and paste it in a corner.
Args:
img (np.ndarray): raw image needed to detailed, format [H, W, C]
cpos_ratio (Tuple[float, float]): selected patch's centroid, from 0 to 1
area_pixels (Tuple[int, int]): pixel area of the patch
interp_ratio (int, optional): interpolate ratio. Defaults to 3.
color (Tuple[int, int, int], optional): color of the box. Defaults to (0, 255, 0).
recommend colors:
(236,229,240)
(233,138,21)
(0,59,54)
thickness (int, optional): thickness of the box. Defaults to 2.
place(str, optional): where to place the interpolated patch.
"""
assert 0 < cpos_ratio[0] < 1 and 0 < cpos_ratio[1] < 1, '@cpos_ratio can only be range (0, 1)'
# to array
if img.ndim == 2:
img = np.repeat(img[..., np.newaxis], 3, axis=-1)
elif img.shape[-1] == 1:
img = np.repeat(img, 3, axis=-1)
img_size = np.array(img.shape[:2])
cpos_pixels = np.array(cpos_ratio) * img_size
cpos_ratio = np.array(cpos_ratio)
area_pixels = np.array(area_pixels)
paste_pixels = area_pixels * interp_ratio
img = img.astype('uint8')
img = np.ascontiguousarray(img)
# bound check
bound = []
for i, j in zip((-1, 1), (-1, 1)):
bound.append([i * area_pixels[0] / 2, j * area_pixels[1] / 2])
bound = np.array(bound)
bound = np.repeat(cpos_pixels[np.newaxis, ...], 2, axis=0) + bound
assert not np.bitwise_or(bound[0] < 0, bound[1] > img_size).any(), \
f'selected range out of image size, image size {img_size} but get selected range {bound}'
# find furthest corner to paste the interpolated patch
if place is None:
furthest_pos_ratio = None
furthest_dis = 0.
for i in (0, 1):
for j in (0, 1):
d = (i - cpos_ratio[0]) ** 2 + (j - cpos_ratio[1]) ** 2
if d > furthest_dis:
furthest_pos_ratio = (i, j)
furthest_dis = d
else:
assert place in ('lt', 'rt', 'lb', 'rb'), '@place should be one of [lt, rt, lb, rb]'
place_dict = {'lt': (0, 0), 'rt': (0, 1), 'lb': (1, 0), 'rb': (1, 1)}
furthest_pos_ratio = place_dict[place]
bound = bound.astype('int')
patch = img[bound[0, 0]:bound[1, 0], bound[0, 1]: bound[1, 1], :]
interp_img = cv2.resize(patch, dsize=paste_pixels[::-1])
box_edge_point = []
cv2.rectangle(img, bound[0][::-1], bound[1][::-1], color, thickness=thickness)
box_pre_thick = thickness // 2
if furthest_pos_ratio == (0, 0):
img[:paste_pixels[0], :paste_pixels[1], :] = interp_img
box_edge_point = [[box_pre_thick, box_pre_thick], paste_pixels[::-1]]
elif furthest_pos_ratio == (1, 0):
img[-paste_pixels[0]:, :paste_pixels[1], :] = interp_img
box_edge_point = [[box_pre_thick, img_size[0] - paste_pixels[0]],
[paste_pixels[1], img_size[0] - box_pre_thick]]
elif furthest_pos_ratio == (0, 1):
img[:paste_pixels[0], -paste_pixels[1]:, :] = interp_img
box_edge_point = [[img_size[1] - paste_pixels[1], box_pre_thick],
[img_size[1] - box_pre_thick, paste_pixels[0]]]
else:
img[-paste_pixels[0]:, -paste_pixels[1]:, :] = interp_img
box_edge_point = [[img_size[1] - paste_pixels[1], img_size[0] - paste_pixels[0]],
[img_size[1] - box_pre_thick, img_size[0] - box_pre_thick]]
cv2.rectangle(img, box_edge_point[0], box_edge_point[1], color, thickness)
return img
def plt_plot_img_without_white_margin(img, *args, **kwargs):
"""
:param img: format [H, W, C]
:param args: plt.imshow args
:param kwargs: plt.imshow kwargs
:return:
"""
width, height = img.shape[:2]
ax = plt.imshow(img, *args, **kwargs)
fig = plt.gcf()
fig.set_size_inches(width / 100, height / 100)
plt.gca().xaxis.set_major_locator(plt.NullLocator())
plt.gca().yaxis.set_major_locator(plt.NullLocator())
plt.subplots_adjust(top=1, bottom=0, left=0, right=1, hspace=0, wspace=0)
plt.margins(0, 0)
plt.gca().set_axis_off()
return fig, ax
if __name__ == '__main__':
import matplotlib.pyplot as plt
import PIL.Image as Image
img = Image.open('../visualized_img/sr/8.eps')
img = np.asarray(img)
img = show_details(img, cpos_ratio=(0.2, 0.8), area_pixels=(50, 50), thickness=2)
plt_plot_img_without_white_margin(img)
plt.show()
|
2977094657/DidaAPI | 1,685 | frontend/docs/api/statistics/general-statistics.md | # 获取通用统计信息
## 接口信息
- **接口路径**: `GET /statistics/general`
- **接口描述**: 获取概览、成就值、趋势等通用统计信息
- **请求方式**: GET
- **认证要求**: 需要先完成微信登录获取认证会话
## 请求参数
无需参数
## 请求示例
```bash
curl -X GET "http://localhost:8000/statistics/general"
```
## 响应格式
### 成功响应
```json
{
"score": 1250,
"level": 5,
"yesterdayCompletedCount": 8,
"todayCompletedCount": 12,
"totalCompletedCount": 1580,
"todayPomoCount": 6,
"yesterdayPomoCount": 4,
"totalPomoCount": 320,
"todayFocusTime": 150,
"yesterdayFocusTime": 100,
"totalFocusTime": 8000,
"targetPomoCount": 8,
"targetFocusTime": 200
}
```
### 错误响应
#### 未认证
```json
{
"error": "no_auth_session",
"message": "未设置认证会话,请先完成微信登录"
}
```
#### 服务错误
```json
{
"error": "service_error",
"message": "获取通用统计信息失败,请稍后重试"
}
```
## 响应字段说明
| 字段名 | 类型 | 描述 |
|--------|------|------|
| score | integer | 成就值 |
| level | integer | 账号等级 |
| yesterdayCompletedCount | integer | 昨日完成任务数 |
| todayCompletedCount | integer | 今日完成任务数 |
| totalCompletedCount | integer | 总完成任务数 |
| todayPomoCount | integer | 今日番茄数 |
| yesterdayPomoCount | integer | 昨日番茄数 |
| totalPomoCount | integer | 总番茄数 |
| todayFocusTime | integer | 今日专注时长(分钟) |
| yesterdayFocusTime | integer | 昨日专注时长(分钟) |
| totalFocusTime | integer | 总专注时长(分钟) |
| targetPomoCount | integer | 目标番茄数 |
| targetFocusTime | integer | 目标专注时长(分钟) |
## 使用说明
1. **认证要求**: 需要先调用微信登录接口获取认证会话
2. **统计范围**: 包含任务完成和番茄专注的综合统计
3. **时间维度**: 提供昨日、今日和总计三个维度的数据
4. **目标对比**: 包含目标设置,便于进度对比
## 相关接口
- [获取用户排名统计](../statistics.md)
- [获取任务统计信息](./task-statistics.md)
- [获取番茄专注概览](../pomodoros.md)
- [微信登录流程](../auth/wechat-login-flow.md)
## 注意事项
- 需要先完成微信登录获取认证会话
- 时间以分钟为单位
- 数据实时更新
- 成就值和等级基于用户活跃度计算
|
2977094657/DidaAPI | 2,950 | frontend/docs/api/tasks/get-trash-tasks.md | # 获取垃圾桶任务
获取垃圾桶中的任务列表。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/project/all/trash/page`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
### Query Parameters
| 参数名 | 类型 | 必填 | 说明 | 示例 |
|--------|------|------|------|------|
| limit | number | 否 | 每页任务数量,默认50 | 50 |
| task_type | number | 否 | 任务类型,默认1 | 1 |
## 响应格式
### 成功响应
返回垃圾桶任务的原始数据:
```json
{
"tasks": [
{
"id": "string",
"projectId": "string",
"sortOrder": 0,
"title": "string",
"startDate": "string",
"dueDate": "string",
"timeZone": "string",
"isFloating": true,
"isAllDay": true,
"reminders": [
{
"id": "string",
"trigger": "string"
}
],
"exDate": [
"string"
],
"priority": 0,
"status": 0,
"items": [
"string"
],
"modifiedTime": "string",
"etag": "string",
"deleted": 0,
"createdTime": "string",
"creator": 0,
"attachments": [
"string"
],
"columnId": "string",
"parentId": "string",
"childIds": [
"string"
],
"kind": "string",
"pinnedTime": "string",
"imgMode": 0,
"deletedBy": 0,
"deletedTime": 0,
"repeatFirstDate": "string",
"repeatTaskId": "string",
"repeatFrom": "string",
"tags": [
"string"
],
"commentCount": 0,
"focusSummaries": [
{
"userId": 0,
"pomoCount": 0,
"estimatedPomo": 0,
"estimatedDuration": 0,
"pomoDuration": 0,
"stopwatchDuration": 0
}
],
"repeatFlag": "string",
"pomodoroSummaries": [
{
"userId": 0,
"count": 0,
"estimatedPomo": 0,
"duration": 0
}
]
}
],
"next": 0
}
```
## 响应字段说明
### 响应结构
| 字段名 | 类型 | 说明 |
|--------|------|------|
| tasks | array | 垃圾桶任务列表 |
| next | number | 下一页标识 |
### 任务对象核心字段
| 字段名 | 类型 | 说明 |
|--------|------|------|
| id | string | 任务唯一标识符 |
| projectId | string | 所属项目ID |
| title | string | 任务标题 |
| status | number | 任务状态 |
| priority | number | 优先级 |
| deleted | number | 删除标志 |
| deletedBy | number | 删除者ID |
| deletedTime | number | 删除时间 |
### 时间字段
| 字段名 | 类型 | 说明 |
|--------|------|------|
| createdTime | string | 创建时间 |
| modifiedTime | string | 修改时间 |
| startDate | string | 开始日期 |
| dueDate | string | 截止日期 |
|
2977094657/DidaAPI | 4,008 | frontend/docs/api/tasks/get-completed-tasks.md | # 获取已完成/已放弃任务
获取用户的已完成或已放弃任务列表,支持分页获取。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/project/all/closed`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
### Query Parameters
| 参数名 | 类型 | 必填 | 说明 | 示例 |
|--------|------|------|------|------|
| to | string | 否 | 分页参数,使用上次响应最后一个任务的completedTime | "2025-03-15T13:30:54.000+0000" |
| status | string | 否 | 任务状态,默认Completed | "Completed" 或 "Abandoned" |
### 状态参数说明
| 状态值 | 说明 |
|--------|------|
| Completed | 已完成的任务(默认) |
| Abandoned | 已放弃的任务 |
### 分页机制
**重要说明**: 此接口的分页机制需要严格按照以下步骤执行,否则无法获取全部数据。
**第一次请求(不携带to字段)**:
```
GET https://api.dida365.com/api/v2/project/all/closed?from=&status=Completed
```
**获取已放弃任务(第一次请求)**:
```
GET https://api.dida365.com/api/v2/project/all/closed?from=&status=Abandoned
```
**后续分页请求**:
```
GET https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Completed
GET https://api.dida365.com/api/v2/project/all/closed?from=&to=2025-03-15%2013:30:54&status=Abandoned
```
**分页流程详解**:
1. **第一次请求**: 不携带`to`字段,直接请求
2. **获取分页参数**: 从响应中取最后一个任务的`completedTime`字段
3. **构造下次请求**: 将`completedTime`作为`to`参数
4. **重复请求**: 直到返回的任务数量少于50个为止
**to参数说明**:
- 使用上次响应最后一个任务的`completedTime`字段值(原始格式)
- 原始格式:`2025-03-15T13:30:54.000+0000`
- API转换格式:`2025-03-15 13:30:54`
- URL编码后:`2025-03-15%2013:30:54`
- **注意**: 传入原始的completedTime格式,API会自动转换为正确格式
**分页终止条件**:
- 当返回的任务数量少于50个时,表示已获取到最后一页
- 当返回空数组时,表示没有更多数据
## 响应格式
### 成功响应
返回已完成任务列表的原始数据:
```json
[
{
"id": "string",
"projectId": "string",
"sortOrder": 0,
"title": "string",
"content": "string",
"desc": "string",
"timeZone": "string",
"isFloating": true,
"isAllDay": true,
"reminder": "string",
"reminders": [
{
"id": "string",
"trigger": "string"
}
],
"repeatFirstDate": "string",
"exDate": [
"string"
],
"completedTime": "string",
"completedUserId": 0,
"repeatTaskId": "string",
"priority": 0,
"status": 0,
"items": [
"string"
],
"progress": 0,
"modifiedTime": "string",
"etag": "string",
"deleted": 0,
"createdTime": "string",
"creator": 0,
"repeatFrom": "string",
"attachments": [
"string"
],
"commentCount": 0,
"focusSummaries": [
{
"userId": 0,
"pomoCount": 0,
"estimatedPomo": 0,
"estimatedDuration": 0,
"pomoDuration": 0,
"stopwatchDuration": 0,
"focuses": [
[
"string"
]
]
}
],
"columnId": "string",
"parentId": "string",
"kind": "string",
"pinnedTime": "string",
"imgMode": 0,
"startDate": "string",
"dueDate": "string",
"tags": [
"string"
],
"pomodoroSummaries": [
{
"userId": 0,
"count": 0,
"estimatedPomo": 0,
"duration": 0
}
],
"repeatFlag": "string",
"childIds": [
"string"
]
}
]
```
## 响应字段说明
### 核心字段
| 字段名 | 类型 | 说明 |
|--------|------|------|
| id | string | 任务唯一标识符 |
| projectId | string | 所属项目ID |
| title | string | 任务标题 |
| content | string | 任务内容 |
| completedTime | string | 完成时间(分页关键字段) |
| priority | number | 优先级 |
| status | number | 任务状态 |
### 时间字段
| 字段名 | 类型 | 说明 |
|--------|------|------|
| completedTime | string | 完成时间 |
| createdTime | string | 创建时间 |
| modifiedTime | string | 修改时间 |
| startDate | string | 开始日期 |
| dueDate | string | 截止日期 |
### 专注统计
| 字段名 | 类型 | 说明 |
|--------|------|------|
| focusSummaries | array | 专注统计摘要 |
| pomodoroSummaries | array | 番茄钟统计摘要 |
|
2977094657/DidaAPI | 11,083 | frontend/docs/api/tasks/get-all-tasks.md | # 获取所有任务
获取当前用户的所有任务列表,包括任务详情、状态、优先级等信息。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/batch/check/0`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 认证要求
此接口需要完整的认证信息,包括:
### 必需的Cookie
```http
Cookie: t={auth_token}; _csrf_token={csrf_token}; AWSALB={load_balancer_cookie}; AWSALBCORS={load_balancer_cookie}
```
### 必需的请求头
```http
Accept: application/json, text/plain, */*
Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7
Cache-Control: no-cache
Origin: https://dida365.com
Pragma: no-cache
Referer: https://dida365.com/
User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36
X-Csrftoken: {csrf_token}
X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":""}
Hl: zh_CN
X-Tz: Asia/Shanghai
Traceid: {trace_id}
```
## 完整请求示例
```http
GET https://api.dida365.com/api/v2/batch/check/0 HTTP/1.1
Host: api.dida365.com
Accept: application/json, text/plain, */*
Accept-Encoding: gzip, deflate, br, zstd
Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7
Cache-Control: no-cache
Cookie: t={auth_token}; _csrf_token={csrf_token}
Hl: zh_CN
Origin: https://dida365.com
Pragma: no-cache
Referer: https://dida365.com/
Traceid: {trace_id}
User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36
X-Csrftoken: {csrf_token}
X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":"{websocket_id}"}
X-Tz: Asia/Shanghai
```
## 响应格式
### 成功响应
**状态码**: `200 OK`
**Content-Type**: `application/json;charset=UTF-8`
### 响应体结构
```json
{
"checkPoint": 0,
"syncTaskBean": {
"update": [
{
"id": "string",
"projectId": "string",
"sortOrder": 0,
"title": "string",
"content": "string",
"desc": "string",
"timeZone": "string",
"isFloating": true,
"isAllDay": true,
"reminder": "string",
"reminders": [
{
"id": "string",
"trigger": "string"
}
],
"exDate": [
"string"
],
"repeatTaskId": "string",
"priority": 0,
"status": 0,
"items": [
"string"
],
"progress": 0,
"modifiedTime": "string",
"etag": "string",
"deleted": 0,
"createdTime": "string",
"creator": 0,
"repeatFrom": "string",
"attachments": [
{
"id": "string",
"refId": "string",
"path": "string",
"size": 0,
"fileName": "string",
"fileType": "string",
"status": 0,
"createdTime": "string"
}
],
"commentCount": 0,
"focusSummaries": [
{
"userId": 0,
"pomoCount": 0,
"estimatedPomo": 0,
"estimatedDuration": 0,
"pomoDuration": 0,
"stopwatchDuration": 0,
"focuses": [
[
"string"
]
]
}
],
"columnId": "string",
"kind": "string",
"imgMode": 0,
"startDate": "string",
"dueDate": "string",
"repeatFlag": "string",
"pomodoroSummaries": [
{
"userId": 0,
"count": 0,
"estimatedPomo": 0,
"duration": 0
}
],
"childIds": [
"string"
],
"pinnedTime": "string",
"repeatFirstDate": "string",
"tags": [
"string"
],
"parentId": "string",
"deletedBy": 0,
"deletedTime": 0
}
],
"tagUpdate": [
"string"
],
"delete": [
"string"
],
"add": [
"string"
],
"empty": true
},
"projectProfiles": [
{
"id": "string",
"name": "string",
"isOwner": true,
"color": "string",
"sortOrder": 0,
"sortOption": {
"groupBy": "string",
"orderBy": "string"
},
"sortType": "string",
"userCount": 0,
"etag": "string",
"modifiedTime": "string",
"inAll": true,
"showType": null,
"muted": true,
"reminderType": null,
"closed": null,
"transferred": null,
"groupId": null,
"viewMode": "string",
"notificationOptions": [
"string"
],
"teamId": null,
"permission": "string",
"kind": "string",
"timeline": {
"range": null,
"sortType": "string",
"sortOption": {
"groupBy": "string",
"orderBy": "string"
}
},
"needAudit": true,
"barcodeNeedAudit": true,
"openToTeam": true,
"teamMemberPermission": null,
"source": 0
}
],
"projectGroups": [
"string"
],
"filters": null,
"tags": [
"string"
],
"syncTaskOrderBean": {
"taskOrderByDate": {},
"taskOrderByPriority": {},
"taskOrderByProject": {}
},
"syncOrderBean": {
"orderByType": {
"taskPinned": {
"{project_id_1}": {
"changed": [
{
"id": "string",
"order": 0,
"type": 0
}
],
"deleted": [
{
"id": "string",
"order": 0,
"type": 0
}
]
},
"{project_id_2}": {
"changed": [
{
"id": "string",
"order": 0,
"type": 0
}
],
"deleted": [
{
"id": "string",
"order": 0,
"type": 0
}
]
}
}
}
},
"syncOrderBeanV3": {
"orderByType": {}
},
"inboxId": "string",
"checks": null,
"remindChanges": [
"string"
]
}
```
## 响应字段说明
### 根级字段
| 字段名 | 类型 | 说明 |
|--------|------|------|
| checkPoint | number | 检查点 |
| syncTaskBean | object | 任务同步数据 |
| projectProfiles | array | 项目配置列表 |
| projectGroups | array | 项目分组 |
| filters | null | 过滤器 |
| tags | array | 标签列表 |
| syncTaskOrderBean | object | 任务排序数据 |
| syncOrderBean | object | 排序数据 |
| syncOrderBeanV3 | object | 排序数据V3 |
| inboxId | string | 收件箱ID |
| checks | null | 检查项 |
| remindChanges | array | 提醒变更 |
### 任务字段 (syncTaskBean.update)
| 字段名 | 类型 | 说明 |
|--------|------|------|
| id | string | 任务唯一标识符 |
| projectId | string | 所属项目ID |
| sortOrder | number | 排序顺序 |
| title | string | 任务标题 |
| content | string | 任务内容 |
| desc | string | 任务描述 |
| timeZone | string | 时区 |
| isFloating | boolean | 是否浮动时间 |
| isAllDay | boolean | 是否全天任务 |
| reminder | string | 提醒设置 |
| reminders | array | 提醒列表 |
| exDate | array | 排除日期 |
| repeatTaskId | string | 重复任务ID |
| priority | number | 优先级 |
| status | number | 任务状态 |
| items | array | 子项目 |
| progress | number | 完成进度 |
| modifiedTime | string | 修改时间 |
| etag | string | 实体标签 |
| deleted | number | 删除状态 |
| createdTime | string | 创建时间 |
| creator | number | 创建者ID |
| repeatFrom | string | 重复来源 |
| attachments | array | 附件列表 |
| commentCount | number | 评论数量 |
| focusSummaries | array | 专注摘要 |
| columnId | string | 列ID |
| kind | string | 类型 |
| imgMode | number | 图片模式 |
| startDate | string | 开始日期 |
| dueDate | string | 截止日期 |
| repeatFlag | string | 重复标志 |
| pomodoroSummaries | array | 番茄钟摘要 |
| childIds | array | 子任务ID列表 |
| pinnedTime | string | 置顶时间 |
| repeatFirstDate | string | 首次重复日期 |
| tags | array | 标签列表 |
| parentId | string | 父任务ID |
| deletedBy | number | 删除者ID |
| deletedTime | number | 删除时间 |
### 项目配置字段 (projectProfiles)
| 字段名 | 类型 | 说明 |
|--------|------|------|
| id | string | 项目ID |
| name | string | 项目名称 |
| isOwner | boolean | 是否为所有者 |
| color | string | 项目颜色 |
| sortOrder | number | 排序顺序 |
| sortOption | object | 排序选项 |
| sortType | string | 排序类型 |
| userCount | number | 用户数量 |
| etag | string | 实体标签 |
| modifiedTime | string | 修改时间 |
| inAll | boolean | 是否在全部中显示 |
| showType | null | 显示类型 |
| muted | boolean | 是否静音 |
| reminderType | null | 提醒类型 |
| closed | null | 是否关闭 |
| transferred | null | 是否转移 |
| groupId | null | 分组ID |
| viewMode | string | 查看模式 |
| notificationOptions | array | 通知选项 |
| teamId | null | 团队ID |
| permission | string | 权限 |
| kind | string | 类型 |
| timeline | object | 时间线配置 |
| needAudit | boolean | 是否需要审核 |
| barcodeNeedAudit | boolean | 条码是否需要审核 |
| openToTeam | boolean | 是否对团队开放 |
| teamMemberPermission | null | 团队成员权限 |
| source | number | 来源 |
### 附件字段 (attachments)
| 字段名 | 类型 | 说明 |
|--------|------|------|
| id | string | 附件ID |
| refId | string | 引用ID |
| path | string | 文件路径 |
| size | number | 文件大小 |
| fileName | string | 文件名 |
| fileType | string | 文件类型 |
| status | number | 状态 |
| createdTime | string | 创建时间 |
### 专注摘要字段 (focusSummaries)
| 字段名 | 类型 | 说明 |
|--------|------|------|
| userId | number | 用户ID |
| pomoCount | number | 番茄钟数量 |
| estimatedPomo | number | 预估番茄钟 |
| estimatedDuration | number | 预估时长 |
| pomoDuration | number | 番茄钟时长 |
| stopwatchDuration | number | 秒表时长 |
| focuses | array | 专注记录 |
## 相关接口
- [验证微信登录](../auth/validate-wechat-login.md) - 获取认证令牌
- [获取已完成任务](./get-completed-tasks.md) - 获取已完成任务列表
- [获取垃圾桶任务](./get-trash-tasks.md) - 获取垃圾桶任务列表
|
2977094657/DidaAPI | 1,276 | frontend/docs/api/tasks/get-tasks-summary.md | # 获取任务统计
## 接口信息
- **接口路径**: `GET /tasks/summary`
- **接口描述**: 获取任务的统计信息
- **请求方式**: GET
- **认证要求**: 需要先完成微信登录获取认证会话
## 请求参数
无需参数
## 请求示例
```bash
curl -X GET "http://localhost:8000/tasks/summary"
```
## 响应格式
### 成功响应
```json
{
"code": 200,
"message": "获取任务统计成功",
"data": {
"total_tasks": 150,
"completed_tasks": 120,
"pending_tasks": 30,
"completion_rate": 80.0
}
}
```
### 错误响应
#### 未认证
```json
{
"error": "获取任务统计失败",
"details": {
"error": "no_auth_session",
"message": "未设置认证会话,请先完成微信登录"
}
}
```
#### 服务错误
```json
{
"detail": "服务器内部错误: 具体错误信息"
}
```
## 响应字段说明
| 字段名 | 类型 | 描述 |
|--------|------|------|
| code | integer | 响应状态码,200表示成功 |
| message | string | 响应消息 |
| data | object | 统计数据对象 |
| data.total_tasks | integer | 总任务数 |
| data.completed_tasks | integer | 已完成任务数 |
| data.pending_tasks | integer | 未完成任务数 |
| data.completion_rate | float | 完成率(百分比) |
## 使用说明
1. **认证要求**: 需要先调用微信登录接口获取认证会话
2. **统计范围**: 统计当前用户的所有任务
3. **实时数据**: 返回实时的任务统计信息
4. **完成率计算**: 完成率 = (已完成任务数 / 总任务数) × 100
## 相关接口
- [获取所有任务](./get-all-tasks.md)
- [获取已完成任务](./get-completed-tasks.md)
- [获取垃圾桶任务](./get-trash-tasks.md)
- [微信登录流程](../auth/wechat-login-flow.md)
## 注意事项
- 需要先完成微信登录获取认证会话
- 统计数据基于用户的所有任务
- 完成率保留两位小数
- 如果没有任务,完成率为0
|
2977094657/DidaAPI | 1,957 | frontend/docs/api/auth/password-login.md | # 密码登录
使用用户名(邮箱或手机号)和密码进行滴答清单登录认证。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/user/signon?wc=true&remember=true`
- **请求方法**: `POST`
- **认证要求**: 无需认证
- **所属平台**: 滴答清单
## 请求参数
### 请求体
| 参数名 | 类型 | 必需 | 说明 |
|----------|--------|------|--------------------------|
| username | string | 是 | 登录账户(邮箱或手机号) |
| password | string | 是 | 登录密码 |
### 请求示例
```json
{
"username": "string",
"password": "string"
}
```
## 响应格式
### 成功响应
**HTTP状态码**: `200 OK`
```json
{
"token": "string",
"userId": "string",
"userCode": "string",
"username": "string",
"teamPro": false,
"proStartDate": "string",
"proEndDate": "string",
"subscribeType": "string",
"subscribeFreq": "string",
"needSubscribe": false,
"freq": "string",
"inboxId": "string",
"teamUser": false,
"activeTeamUser": false,
"freeTrial": false,
"gracePeriod": false,
"pro": true,
"ds": false
}
```
## 响应字段说明
| 字段名 | 类型 | 说明 |
|------------------|---------|------------------------|
| token | string | 认证令牌,用于后续API调用 |
| userId | string | 用户唯一ID |
| userCode | string | 用户代码 |
| username | string | 用户名(邮箱或手机号) |
| teamPro | boolean | 是否为团队专业版 |
| proStartDate | string | 专业版开始日期 |
| proEndDate | string | 专业版结束日期 |
| subscribeType | string | 订阅类型(如:wxpay_subscribe) |
| subscribeFreq | string | 订阅频率(如:Month) |
| needSubscribe | boolean | 是否需要订阅 |
| freq | string | 频率 |
| inboxId | string | 默认任务添加清单ID |
| teamUser | boolean | 是否为团队用户 |
| activeTeamUser | boolean | 是否为活跃团队用户 |
| freeTrial | boolean | 是否为免费试用 |
| gracePeriod | boolean | 是否在宽限期 |
| pro | boolean | 是否为专业版 |
| ds | boolean | 数据同步状态 | |
281677160/openwrt-package | 1,077 | luci-app-passwall/root/usr/share/passwall/lease2hosts.sh | #!/bin/sh
# dhcp.leases to hosts
CONFIG=passwall
TMP_PATH=/tmp/etc/${CONFIG}
TMP_PATH2=/tmp/etc/${CONFIG}_tmp
LOCK_FILE=/tmp/lock/${CONFIG}_lease2hosts.lock
LEASE_FILE="/tmp/dhcp.leases"
HOSTS_FILE="$TMP_PATH2/dhcp-hosts"
TMP_FILE="/tmp/dhcp-hosts.tmp"
exec 99>"$LOCK_FILE"
flock -n 99
if [ "$?" != 0 ]; then
exit 0
fi
reload_dnsmasq_pids() {
local pidfile pid
find $TMP_PATH/acl -type f -name 'dnsmasq.pid' 2>/dev/null | while read pidfile; do
if [ -s "$pidfile" ]; then
read pid < "$pidfile"
if [ -n "$pid" ] && kill -0 "$pid" 2>/dev/null; then
kill -HUP "$pid"
fi
fi
done
}
while true; do
if [ -f "$LEASE_FILE" ]; then
awk 'NF >= 4 && $4 != "*" {print $3" "$4}' "$LEASE_FILE" | sort > "$TMP_FILE"
if [ -s "$TMP_FILE" ]; then
if [ ! -f "$HOSTS_FILE" ] || ! cmp -s "$TMP_FILE" "$HOSTS_FILE"; then
mv "$TMP_FILE" "$HOSTS_FILE"
reload_dnsmasq_pids
else
rm -f "$TMP_FILE"
fi
else
if [ -s "$HOSTS_FILE" ]; then
: > "$HOSTS_FILE"
reload_dnsmasq_pids
fi
rm -f "$TMP_FILE"
fi
fi
sleep 60
done 2>/dev/null
|
294coder/Efficient-MIF | 14,989 | utils/_metric_legacy.py | #####
# copy from https://github.com/XiaoXiao-Woo/PanCollection/blob/dev/UDL/pansharpening/common/evaluate.py
# thanks a lot
#####
import math
import torch
import torch.nn.functional as F
import numpy as np
from torch.autograd import Variable
import decimal
# 由于dat及其方差等数值舍入存在误差,最终结果有0.001左右的误差
def q2n(gt, x, q_blocks_size, q_shift):
'''
'''
if isinstance(gt, torch.Tensor):
gt = gt.cpu().numpy()
x = x.cpu().numpy()
N, N1, N2, N3 = gt.shape # 255 255 8
size2 = q_blocks_size # 32
stepx = math.ceil(N1 / q_shift) # 8
stepy = math.ceil(N2 / q_shift) # 8
if stepy <= 0:
stepy = 1
stepx = 1
est1 = (stepx - 1) * q_shift + q_blocks_size - N1 # 1
est2 = (stepy - 1) * q_shift + q_blocks_size - N2 # 1
# if np.sum(np.array([est1 != 0, est2 != 0])) > 0:
# refref = np.zeros(shape=[N1+1, N2+1])
# fusfus = refref.copy()
for i in range(N3):
a1 = gt[..., 0]
ia1 = np.zeros(shape=[N, N1 + est1, N2 + est2])
ia1[:, : N1, : N2] = a1
ia1[:, :, N2:N2 + est2] = ia1[:, :, N2 - 1:-1:N2 - est2 + 1]
ia1[:, N1:N1 + est1, ...] = ia1[:, N1 - 1:-1:N1 - est1 + 1, ...]
if i == 0:
refref = ia1[..., np.newaxis] # np.concatenate(refref, ia1, axis=3)
else:
refref = np.concatenate([refref, ia1[..., np.newaxis]], axis=-1)
if i < N3:
gt = gt[..., 1:]
gt = refref
for i in range(N3):
a2 = x[..., 0]
ia2 = np.zeros(shape=[N, N1 + est1, N2 + est2])
ia2[:, : N1, : N2] = a2
ia2[:, :, N2:N2 + est2] = ia2[:, :, N2 - 1:-1:N2 - est2 + 1]
ia2[:, N1:N1 + est1, ...] = ia2[:, N1 - 1:-1:N1 - est1 + 1, ...]
if i == 0:
fusfus = ia2[..., np.newaxis] # np.concatenate(refref, ia1, axis=3)
else:
fusfus = np.concatenate([fusfus, ia2[..., np.newaxis]], axis=-1)
if i < N3:
x = x[..., 1:]
x = fusfus
x = np.array(x, dtype=np.uint16)
gt = np.array(gt, dtype=np.uint16)
_, N1, N2, N3 = gt.shape
if math.ceil(math.log2(N3)) - math.log2(N3) != 0:
Ndif = pow(2, math.ceil(math.log2(N3))) - N3
dif = np.zeros(shape=[N, N1, N2, Ndif], dtype=np.uint16)
gt = np.concatenate(gt, dif, axis=-1)
x = np.concatenate(x, dif, axis=-1)
_, _, _, N3 = gt.shape
valori = np.zeros(shape=[N, stepx, stepy, N3])
for j in range(stepx):
for i in range(stepy):
o = onions_quality(gt[:, j * q_shift:j * q_shift + q_blocks_size,
i * q_shift: i * q_shift + size2, :],
x[:, j * q_shift:j * q_shift + q_blocks_size,
i * q_shift: i * q_shift + size2, :],
q_blocks_size)
# 0.971379489438014 0.00553590637316723 0.00305237797490489 -0.0188289323262161 -0.00420556598390016 -0.0173947468044076 -0.0202144450367593 0.0102693855205061
valori[:, j, i, :] = o
q2n_idx_map = np.sqrt(np.sum(valori ** 2, axis=-1))
# q2n_index = np.mean(q2n_idx_map)
return q2n_idx_map
def norm_blocco(x, eps=1e-8):
a = x.mean()
c = x.std()
if c == 0:
c = eps
return (x - a) / c + 1, a, c
def onions_quality(dat1, dat2, size1):
dat1 = np.float64(dat1)
dat2 = np.float64(dat2)
dat2 = np.concatenate([dat2[..., 0, np.newaxis], -dat2[..., 1:]], axis=-1)
N, _, _, N3 = dat1.shape
size2 = size1
# Block norm
'''
319.6474609375 37.05174450544686
357.970703125 61.54042371537683
518.708984375 111.53732768071865
608.23828125 154.26606056030568
593.412109375 163.97722215177643
554.8486328125 113.96758695803403
690.16015625 151.29524031046248
442.2314453125 94.12877724960003
mat
319.6475 37.0698
357.9707 61.5705
518.7090 111.5918
608.2383 154.3414
593.4121 164.0573
554.8486 114.0233
690.1602 151.3692
442.2314 94.1748
'''
for i in range(N3):
a1, s, t = norm_blocco(np.squeeze(dat1[..., i]))
# print(s,t)
dat1[..., i] = a1
if s == 0:
if i == 0:
dat2[..., i] = dat2[..., i] - s + 1
else:
dat2[..., i] = -(-dat2[..., i] - s + 1)
else:
if i == 0:
dat2[..., i] = ((dat2[..., i] - s) / t) + 1
else:
dat2[..., i] = -(((-dat2[..., i] - s) / t) + 1)
m1 = np.zeros(shape=[N, N3])
m2 = m1.copy()
mod_q1m = 0
mod_q2m = 0
mod_q1 = np.zeros(shape=[size1, size2])
mod_q2 = np.zeros(shape=[size1, size2])
for i in range(N3):
m1[..., i] = np.mean(np.squeeze(dat1[..., i]))
m2[..., i] = np.mean(np.squeeze(dat2[..., i]))
mod_q1m += m1[..., i] ** 2
mod_q2m += m2[..., i] ** 2
mod_q1 += np.squeeze(dat1[..., i]) ** 2
mod_q2 += np.squeeze(dat2[..., i]) ** 2
mod_q1m = np.sqrt(mod_q1m)
mod_q2m = np.sqrt(mod_q2m)
mod_q1 = np.sqrt(mod_q1)
mod_q2 = np.sqrt(mod_q2)
termine2 = mod_q1m * mod_q2m # 7.97
termine4 = mod_q1m ** 2 + mod_q2m ** 2 #
int1 = (size1 * size2) / (size1 * size2 - 1) * np.mean(mod_q1 ** 2)
int2 = (size1 * size2) / (size1 * size2 - 1) * np.mean(mod_q2 ** 2)
termine3 = int1 + int2 - (size1 * size2) / ((size1 * size2 - 1)) * (mod_q1m ** 2 + mod_q2m ** 2) # 17.8988 ** 2
mean_bias = 2 * termine2 / termine4 # 1
if termine3 == 0:
q = np.zeros(shape=[N, 1, N3])
q[:, :, N3 - 1] = mean_bias
else:
cbm = 2 / termine3
# 32 32 8
qu = onion_mult2D(dat1, dat2)
qm = onion_mult(m1.reshape(-1), m2.reshape(-1))
qv = np.zeros(shape=[N, N3])
for i in range(N3):
qv[..., i] = (size1 * size2) / ((size1 * size2) - 1) * np.mean(np.squeeze(qu[:, :, i]))
q = qv - (size1 * size2) / ((size1 * size2) - 1) * qm
q = q * mean_bias * cbm
return q
def onion_mult2D(onion1, onion2):
_, _, _, N3 = onion1.shape
if N3 > 1:
L = N3 // 2
a = onion1[..., : L]
b = onion1[..., L:]
b = np.concatenate([b[..., 0, np.newaxis], -b[..., 1:]], axis=-1)
c = onion2[..., : L]
d = onion2[..., L:]
d = np.concatenate([d[..., 0, np.newaxis], -d[..., 1:]], axis=-1)
if N3 == 2:
ris = np.concatenate([a * c - d * b, a * d + c * b], axis=-1)
else:
ris1 = onion_mult2D(a, c)
ris2 = onion_mult2D(d, np.concatenate([b[..., 0, np.newaxis], -b[..., 1:]], axis=-1))
ris3 = onion_mult2D(np.concatenate([a[..., 0, np.newaxis], -a[..., 1:]], axis=-1), d)
ris4 = onion_mult2D(c, b)
aux1 = ris1 - ris2
aux2 = ris3 + ris4
ris = np.concatenate([aux1, aux2], axis=-1)
else:
ris = onion1 * onion2
return ris
def onion_mult(onion1, onion2):
# _, N = onion1.shape
N = len(onion1)
if N > 1:
L = N // 2
a = onion1[:L]
b = onion1[L:]
# b[1:] = -b[1:]
b = np.append(np.array(b[0]), -b[1:])
c = onion2[:L]
d = onion2[L:]
# d[1:] = -d[1:]
d = np.append(np.array(d[0]), -d[1:])
if N == 2:
ris = np.append(a * c - d * b, a * d + c * b)
else:
ris1 = onion_mult(a, c)
# b[1:] = -b[1:]
ris2 = onion_mult(d, np.append(np.array(b[0]), -b[1:]))
# a[1:] = -a[1:]
ris3 = onion_mult(np.append(np.array(a[0]), -a[1:]), d)
ris4 = onion_mult(c, b)
aux1 = ris1 - ris2
aux2 = ris3 + ris4
ris = np.append(aux1, aux2)
else:
ris = np.array(onion1).reshape(-1) * np.array(onion2).reshape(-1)
return ris
def compute_index(img_base, img_out, ratio):
h = img_out.shape[0]
w = img_out.shape[1]
chanel = img_out.shape[2]
# 计算SAM
sum1 = torch.sum(img_base * img_out, 2)
sum2 = torch.sum(img_base * img_base, 2)
sum3 = torch.sum(img_out * img_out, 2)
t = (sum2 * sum3) ** 0.5
numlocal = torch.gt(t, 0)
num = torch.sum(numlocal)
t = sum1 / t
angle = torch.acos(t)
sumangle = torch.where(torch.isnan(angle), torch.full_like(angle, 0), angle).sum()
if num == 0:
averangle = sumangle
else:
averangle = sumangle / num
SAM = averangle * 180 / 3.14159256
# 计算ERGAS
summ = 0
for i in range(chanel):
a1 = torch.mean((img_base[:, :, i] - img_out[:, :, i]) ** 2)
m1 = torch.mean(img_base[:, :, i])
a2 = m1 * m1
summ = summ + a1 / a2
ERGAS = 100 * (1 / ratio) * ((summ / chanel) ** 0.5)
return SAM, ERGAS
decimal.getcontext().rounding = "ROUND_HALF_UP"
n_digits = 6
# panHrnet: 2.6565 |1.4651 | 0.98364 | 0.98024 | 0.98089-Q8
def analysis_accu(img_base, img_out, ratio, flag_cut_bounds=True, dim_cut=1, choices=4):
if flag_cut_bounds:
img_base = img_base[dim_cut - 1:-dim_cut, dim_cut - 1:-dim_cut, :] #:
img_out = img_out[dim_cut - 1:-dim_cut, dim_cut - 1:-dim_cut, :] #:
# q2n
# q2n_index = q2n(img_base, img_out, q_blocks_size=32, q_shift=32)
h = img_out.shape[0]
w = img_out.shape[1]
chanel = img_out.shape[2]
# 计算SAM
sum1 = torch.sum(img_base * img_out, 2)
sum2 = torch.sum(img_base * img_base, 2)
sum3 = torch.sum(img_out * img_out, 2)
t = (sum2 * sum3) ** 0.5
numlocal = torch.gt(t, 0)
num = torch.sum(numlocal)
t = sum1 / t
angle = torch.acos(t)
sumangle = torch.where(torch.isnan(angle), torch.full_like(angle, 0), angle).sum()
if num == 0:
averangle = sumangle
else:
averangle = sumangle / num
# 或者采用https://segmentfault.com/a/1190000018929994修改精度
# averangle = math.ceil(averangle * 1000000) / 1000000
averangle = (averangle * 10 ** n_digits).round() / (10 ** n_digits)
# SAM = decimal.Decimal(averangle.cpu().numpy() * 180 / 3.14159256).quantize(decimal.Decimal("0.00000"))
SAM = averangle * 180 / 3.14159256
# 计算ERGAS
summ = 0
for i in range(chanel):
a1 = torch.mean((img_base[:, :, i] - img_out[:, :, i]) ** 2)
m1 = torch.mean(img_base[:, :, i])
a2 = m1 * m1
summ = summ + a1 / a2
ERGAS = 100 * (1 / ratio) * ((summ / chanel) ** 0.5)
# 计算PSNR
mse = torch.mean((img_base - img_out) ** 2, 0)
mse = torch.mean(mse, 0)
rmse = mse ** 0.5
temp = torch.log(1 / rmse) / math.log(10)
PSNR = 20 * temp
# 计算SSIM
# img_base = img_base.permute(2, 0, 1)
# img_out = img_out.permute(2, 0, 1)
# img_base = img_base.unsqueeze(0)
# img_out = img_out.unsqueeze(0)
# SSIM = _ssim(img_base.permute(2, 0, 1).unsqueeze(0), img_out.permute(2, 0, 1).unsqueeze(0))
# index = torch.zeros((5, chanel + 1))
# index[0, 1:chanel + 1] = CC
# index[1, 1:chanel + 1] = PSNR
# index[2, 1:chanel + 1] = SSIM
# index[0, 0] = torch.mean(CC)
# index[1, 0] = torch.mean(PSNR)
# index[2, 0] = torch.mean(SSIM)
# index[3, 0] = SAM
# index[4, 0] = ERGAS
PSNR = torch.mean(PSNR)
# SSIM = torch.mean(SSIM)
# q2n_index = np.mean(q2n_index)
if choices == 5:
# 计算CC
C1 = torch.sum(torch.sum(img_base * img_out, 0), 0) - h * w * (
torch.mean(torch.mean(img_base, 0), 0) * torch.mean(torch.mean(img_out, 0), 0))
C2 = torch.sum(torch.sum(img_out ** 2, 0), 0) - h * w * (torch.mean(torch.mean(img_out, 0), 0) ** 2)
C3 = torch.sum(torch.sum(img_base ** 2, 0), 0) - h * w * (torch.mean(torch.mean(img_base, 0), 0) ** 2)
CC = C1 / ((C2 * C3) ** 0.5)
CC = torch.mean(CC)
return {'SAM': SAM, 'ERGAS': ERGAS, 'PSNR': PSNR, 'CC': CC} # , q2n_index
return {'SAM': SAM, 'ERGAS': ERGAS, 'PSNR': PSNR, }
def _ssim(img1, img2):
img1 = img1.float()
img2 = img2.float()
channel = img1.shape[1]
max_val = 1
_, c, w, h = img1.size()
window_size = min(w, h, 11)
sigma = 1.5 * window_size / 11
window = create_window(window_size, sigma, channel).cuda()
mu1 = F.conv2d(img1, window, padding=window_size // 2, groups=channel)
mu2 = F.conv2d(img2, window, padding=window_size // 2, groups=channel)
mu1_sq = mu1.pow(2)
mu2_sq = mu2.pow(2)
mu1_mu2 = mu1 * mu2
sigma1_sq = F.conv2d(img1 * img1, window, padding=window_size // 2, groups=channel) - mu1_sq
sigma2_sq = F.conv2d(img2 * img2, window, padding=window_size // 2, groups=channel) - mu2_sq
sigma12 = F.conv2d(img1 * img2, window, padding=window_size // 2, groups=channel) - mu1_mu2
C1 = (0.01 * max_val) ** 2
C2 = (0.03 * max_val) ** 2
V1 = 2.0 * sigma12 + C2
V2 = sigma1_sq + sigma2_sq + C2
ssim_map = ((2 * mu1_mu2 + C1) * V1) / ((mu1_sq + mu2_sq + C1) * V2)
t = ssim_map.shape
return ssim_map.mean(2).mean(2)
def gaussian(window_size, sigma):
gauss = torch.Tensor([math.exp(-(x - window_size // 2) ** 2 / float(2 * sigma ** 2)) for x in range(window_size)])
return gauss / gauss.sum()
def create_window(window_size, sigma, channel):
_1D_window = gaussian(window_size, sigma).unsqueeze(1)
_2D_window = _1D_window.mm(_1D_window.t()).float().unsqueeze(0).unsqueeze(0)
window = Variable(_2D_window.expand(channel, 1, window_size, window_size).contiguous())
return window
def compare_index(A):
A_size = A.shape
ite_n = A_size[2]
band_n = A_size[1]
C_better = A[:, 0, 0]
ind = 0
for i in range(ite_n):
score_b = 0
score_c = 0
C_compare = A[:, 0, i]
if (C_better[0] > C_compare[0]):
score_b = score_b + 1
else:
score_c = score_c + 1
if (C_better[1] > C_compare[1]):
score_b = score_b + 1
else:
score_c = score_c + 1
if (C_better[2] > C_compare[2]):
score_b = score_b + 1
else:
score_c = score_c + 1
if (C_better[3] < C_compare[3]):
score_b = score_b + 1
else:
score_c = score_c + 1
if (C_better[4] < C_compare[4]):
score_b = score_b + 1
else:
score_c = score_c + 1
if (score_c > score_b):
C_better = A[:, 0, i]
ind = i
C_best = A[:, :, ind]
best_ind = ind + 1
return C_best, best_ind
########################################
# Non-reference quality assessment
# D_\lambda, D_s, HQNR implementaion
# Author: Xiao Wu, Zihan Cao
# Date: 2024/02/04
########################################
from .sharpening_index.hqnr2 import indexes_evaluation_FS
if __name__ == "__main__":
a = torch.randn(256, 256, 3)
b = torch.randn(256, 256, 3)
print(analysis_accu(a, b, 1, choices=5))
|
294coder/Efficient-MIF | 9,158 | utils/network_utils.py | import math
from typing import Union
from copy import deepcopy
from bytecode import Bytecode, Instr
from accelerate import Accelerator
from accelerate.utils import DistributedType
import torch
import torch.nn as nn
def hook_model(model: nn.Module, saved_tensor, hook_class):
def feature_hook(_, input, output):
# forward hook
saved_tensor.append([input, output])
hooks = []
for n, m in model.named_modules():
if isinstance(m, hook_class):
hooks.append(m.register_forward_hook(feature_hook))
return model, hooks
class loss_with_l2_regularization(nn.Module):
def __init__(self):
super(loss_with_l2_regularization, self).__init__()
def forward(self, loss, model, weight_decay=1e-5, flag=False):
regularizations = []
for k, v in model.named_parameters():
if "conv" in k and "weight" in k:
# print(k)
penality = weight_decay * ((v.data ** 2).sum() / 2)
regularizations.append(penality)
if flag:
print("{} : {}".format(k, penality))
# r = torch.sum(regularizations)
loss = loss + sum(regularizations)
return loss
def variance_scaling_initializer(tensor):
# stole it from woo-xiao.
# thanks
def calculate_fan(shape, factor=2.0, mode="FAN_IN", uniform=False):
# 64 9 3 3 -> 3 3 9 64
# 64 64 3 3 -> 3 3 64 64
if shape:
# fan_in = float(shape[1]) if len(shape) > 1 else float(shape[0])
# fan_out = float(shape[0])
fan_in = float(shape[-2]) if len(shape) > 1 else float(shape[-1])
fan_out = float(shape[-1])
else:
fan_in = 1.0
fan_out = 1.0
for dim in shape[:-2]:
fan_in *= float(dim)
fan_out *= float(dim)
if mode == "FAN_IN":
# Count only number of input connections.
n = fan_in
elif mode == "FAN_OUT":
# Count only number of output connections.
n = fan_out
elif mode == "FAN_AVG":
# Average number of inputs and output connections.
n = (fan_in + fan_out) / 2.0
if uniform:
raise NotImplemented
# # To get stddev = math.sqrt(factor / n) need to adjust for uniform.
# limit = math.sqrt(3.0 * factor / n)
# return random_ops.random_uniform(shape, -limit, limit,
# dtype, seed=seed)
else:
# To get stddev = math.sqrt(factor / n) need to adjust for truncated.
trunc_stddev = math.sqrt(1.3 * factor / n)
return fan_in, fan_out, trunc_stddev
def model_params(model: nn.Module, accelerator=None):
if accelerator is not None:
model = accelerator.unwrap_model(model)
if accelerator.distributed_type == DistributedType.DEEPSPEED:
from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint
return get_fp32_state_dict_from_zero_checkpoint(model)
else:
return model.state_dict()
if isinstance(model, (nn.DataParallel, nn.parallel.DistributedDataParallel)):
model = model.module
elif isinstance(model, torch._dynamo.eval_frame.OptimizedModule): # torch.compile model
model = model._orig_mod
return model.state_dict()
def model_device(model: Union[nn.Module, nn.DataParallel,
nn.parallel.DistributedDataParallel,
torch._dynamo.eval_frame.OptimizedModule]):
params = model.parameters()
p0 = next(params)
return p0.device
def clip_norm_(max_norm, network, fp_scaler=None, optim=None, accelerator=None):
if fp_scaler is not None:
fp_scaler.unscale_(optim)
if accelerator is not None:
accelerator.clip_grad_norm_(network.parameters(), max_norm)
else:
torch.nn.utils.clip_grad.clip_grad_norm_(network.parameters(), max_norm)
def clip_value_(max_value, network, fp_scaler=None, optim=None, accelerator=None):
if fp_scaler is not None:
fp_scaler.unscale_(optim)
if accelerator is not None:
accelerator.clip_grad_value(network.parameters(), max_value)
else:
torch.nn.utils.clip_grad.clip_grad_value_(network.parameters(), max_value)
def step_loss_backward(
optim,
network=None,
max_norm=None,
max_value=None,
loss=None,
fp16=False, # decrepted
mix_precision=False,
fp_scaler=None,
grad_accum=False,
accelerator: Accelerator=None,
):
"""
:param optim: optimizer. type: optim.Optimizer
:param network: instanced network. type: nn.Module
:param max_norm: clip norm. type: float
:param loss: float
:param fp16: bool
:param fp_scaler: mix-precision scaler
:return:
"""
if (fp16 and fp_scaler is None) or (not fp16 and fp_scaler is not None):
if accelerator is None:
raise ValueError("fp16 and grad_scaler should be set together")
else:
fp16 = False
if max_norm is not None and network is None:
raise ValueError("max_norm is set, network should be set")
mixed_precision = fp16
optim.zero_grad()
if mixed_precision:
fp_scaler.scale(loss).backward()
if max_norm is not None:
clip_norm_(max_norm, network, fp_scaler, optim)
if not grad_accum:
fp_scaler.step(optim)
fp_scaler.update()
else:
if accelerator is None:
loss.backward()
else:
accelerator.backward(loss)
# assert max_norm and max_value can not be set at the same time
if max_norm is not None:
clip_norm_(max_norm, network, accelerator=accelerator)
elif max_value is not None:
clip_value_(network.parameters(), max_value, accelerator=accelerator)
if not grad_accum:
optim.step()
def find_no_or_big_norm_params(net: nn.Module, ktop: int=20):
"""find unused params and big-normed gradient
Args:
net (nn.Module): network to be checked.
ktop (int, optional): top k params to be printed. Defaults to 20.
"""
# find unused params and big-normed gradient
d_grads = {}
n_params = 0
for n, p in net.named_parameters():
n_params += p.numel()
if p.grad is None:
print(n, "has no grad")
else:
p_sum = torch.abs(p.grad).sum().item()
d_grads[n] = p_sum
# topk
d_grads = dict(sorted(d_grads.items(), key=lambda item: item[1], reverse=True))
for k, v in list(d_grads.items())[:ktop]:
print(k, v)
class EMAModel(object):
def __init__(self, model, ema_ratio=0.9999):
super().__init__()
self.model = model
self.ema_ratio = ema_ratio
self.ema_model = deepcopy(model)
def update(self):
for ema_p, now_p in zip(self.ema_model.state_dict(), self.model.state_dict()):
ema_p.data = ema_p.data * self.ema_ratio + now_p.data * (1 - self.ema_ratio)
def ema_model_state_dict(self):
try:
return self.ema_model.module.state_dict()
except:
return self.ema_model.state_dict()
class get_local(object):
cache = {}
is_activate = False
def __init__(self, *args):
self.varname = args
def __call__(self, func):
if not type(self).is_activate:
return func
type(self).cache[func.__qualname__] = []
c = Bytecode.from_code(func.__code__)
extra_code = []
extra_code.extend([
*[Instr('LOAD_FAST', varn) for varn in self.varname],
Instr('BUILD_LIST', len(self.varname)),
Instr('STORE_FAST', '_result_list'),
Instr('LOAD_FAST', '_result_list'),
Instr('BUILD_TUPLE', 2)
])
c[-1:-1] = extra_code
func.__code__ = c.to_code()
def wrapper(*args, **kwargs):
res = func(*args, **kwargs)
output_vs = res[0]
saved_vs = res[1]
# assume is a Tensor
for i, v in enumerate(saved_vs):
if hasattr(v, 'detach'):
v = v.detach().cpu() #.numpy()
saved_vs[i] = v
type(self).cache[func.__qualname__].append(saved_vs)
return output_vs
return wrapper
@classmethod
def clear(cls):
for key in cls.cache.keys():
cls.cache[key] = []
@classmethod
def activate(cls):
from loguru import logger
logger.warning('ready to get local varibles, be careful about you CPU memory.')
cls.is_activate = True
# if __name__ == '__main__':
# get_local.activate()
# @get_local('x', 'y')
# def func():
# x = 1
# y = [1,2]
# return x, y
# print(func())
# print(get_local.cache)
|
294coder/Efficient-MIF | 10,596 | utils/metric.py | import numpy as np
import torch
from skimage.metrics import structural_similarity, peak_signal_noise_ratio
from functools import partial
from warnings import warn
import sys
sys.path.append('./')
from utils.misc import to_numpy
from utils.misc import dict_to_str, to_numpy
from utils._metric_legacy import analysis_accu, indexes_evaluation_FS
class NonAnalysis(object):
def __init__(self):
self.acc_ave = {} # only used as attribution
pass
def __call__(self, *args, **kwargs):
pass
def __repr__(self):
return 'NonAnalysis()'
# TODO: need to be tested the new metric analysis
# FIXME: this python code is not same as matlab code, you should use matlab code to get the real accuracy
# only used in training and validate
class AnalysisPanAcc(object):
def __init__(self, ratio=4, ref=True, ergas_ratio: int = 4, **unref_factory_kwargs):
"""pansharpening metric analysis class
Args:
ratio (int, optional): fusion ratio. Defaults to 4.
ref (bool, optional): reduce-resolution or full-resolution. Defaults to True.
ergas_ratio (int, optional): previous api (may decrepated soon). Defaults to 4.
unref_factory_kwargs(dict): sensor, default_max_value. Defaults to {'sensor': 'default', 'default_max_value': None}.
Raises:
ValueError: _description_
"""
# ergas_ratio is decrepated
if ratio is None:
ratio = ergas_ratio
warn('@ergas_ratio is deprecated, use ratio instead')
self.ratio = ratio
self.ref = ref
# metric functions
if ref:
self.__sam_ergas_psnr_cc_one_image = partial(analysis_accu, ratio=ergas_ratio, choices=5)
self.ssim = ssim_batch_tensor_metric
else:
# @sensor in ['QB', 'IKONOS', 'WV2', 'WV3', 'default']
assert 'sensor' in unref_factory_kwargs or 'default_max_value' in unref_factory_kwargs, \
'@sensor or @default_max_value should be specified in unrefactory_kwargs'
sensor = unref_factory_kwargs.pop('sensor', 'default').upper()
if sensor == 'DEFAULT': warn('sensor is not specified, use default sensor type')
self.default_max_value = unref_factory_kwargs.pop('default_max_value', None)
if self.default_max_value is None:
_default_max_value = {'QB': 2047, 'IKONOS': 1023, 'WV2': 2047, 'WV3': 2047,
'GF2': 1023, 'DEFAULT': 2047,
'CAVE_X4': 1, 'CAVE_X8': 1, 'HARVARD_X': 1, 'HARVARD_X8': 1,
'GF5':1, 'GF2-GF5': 1,}
self.default_max_value = _default_max_value.get(sensor)
print(f'>>> @default_max_value is not specified, set it according to @sensor:'
f'{sensor, self.default_max_value}\n', '-'*20)
self.FS_metric_fn = partial(indexes_evaluation_FS, L=11, Qblocks_size=32, sensor=sensor,
th_values=0, ratio=ratio, flagQNR=False)
# tracking accuracy
self._acc_d = {}
self._call_n = 0
self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if ref else \
{'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
@property
def empty_acc(self):
return {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \
{'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
@staticmethod
def permute_dim(*args, permute_dims=(1, 2, 0)):
l = []
for i in args:
l.append(i.permute(*permute_dims))
return l
@staticmethod
def _sum_acc(d_ave, d_now, n, n2=1):
assert len(d_ave) == len(d_now)
for k in d_ave.keys():
v2 = d_now[k] * n2
d_ave[k] *= n
d_ave[k] += v2.cpu().item() if isinstance(v2, torch.Tensor) else v2
return d_ave
@staticmethod
def _average_acc(d_ave, n):
for k in d_ave.keys():
d_ave[k] /= n
return d_ave
def sam_ergas_psnr_cc_batch(self, b_gt, b_pred):
n = b_gt.shape[0]
# input shape should be [B, C, H, W]
acc_ds = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0.}
for i, (img1, img2) in enumerate(zip(b_gt, b_pred)):
img1, img2 = self.permute_dim(img1, img2)
acc_d = self.__sam_ergas_psnr_cc_one_image(img1, img2)
acc_ds = self._sum_acc(acc_ds, acc_d, i)
acc_ds = self._average_acc(acc_ds, i + 1)
return acc_ds
def D_lambda_D_s_HQNR_batch(self, sr=None, ms=None, lms=None, pan=None):
assert sr is not None and lms is not None and pan is not None and ms is not None
if ms is None:
ms = torch.nn.functional.interpolate(lms, scale_factor=1/self.rato, mode='bilinear', align_corners=False)
acc_ds = {'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
sr, ms, lms, pan = self.permute_dim(sr, ms, lms, pan, permute_dims=(0, 2, 3, 1))
sr, ms, lms, pan = to_numpy(sr, ms, lms, pan)
_max_value = getattr(self, 'default_max_value')
sr, ms, lms, pan = map(lambda x: np.clip(x * _max_value, 0, _max_value), [sr, ms, lms, pan])
for i, (sr_i, ms_i, lms_i, pan_i) in enumerate(zip(sr, ms, lms, pan)):
QNR_index, D_lambda, D_S, _ = self.FS_metric_fn(I_F=sr_i, I_MS_LR=ms_i, I_MS=lms_i, I_PAN=pan_i)
acc_d = dict(HQNR=QNR_index, D_lambda=D_lambda, D_S=D_S)
acc_ds = self._sum_acc(acc_ds, acc_d, i)
acc_ds = self._average_acc(acc_ds, i + 1)
return acc_ds
def once_batch_call(self, **kwargs):
if self.ref:
acc_d1 = self.sam_ergas_psnr_cc_batch(**kwargs)
acc_ssim = self.ssim(**kwargs)
acc_d1['SSIM'] = acc_ssim
else:
acc_d1 = self.D_lambda_D_s_HQNR_batch(**kwargs)
self._acc_d = acc_d1
return acc_d1
def _call_check_args_to_kwargs(self, *args):
def may_np_to_tensor(d):
for k, v in d.items():
if not isinstance(v, torch.Tensor):
d[k] = torch.tensor(v, dtype=torch.float32)
elif v.dtype != torch.float32:
d[k] = v.float()
return d
if len(args) == 2:
assert self.ref, 'ref mode should have 2 args'
kwargs = dict(b_gt=args[0], b_pred=args[1])
elif len(args) == 3:
assert not self.ref, 'unref mode should have more than 2 args'
kwargs = dict(sr=args[0], lms=args[1], pan=args[2])
elif len(args) == 4:
assert not self.ref, 'unref mode should have more than 2 args'
kwargs = dict(sr=args[0], ms=args[1], lms=args[2], pan=args[3])
else:
raise ValueError('args should have 2 or 4 elements')
return may_np_to_tensor(kwargs)
def __call__(self, *args):
"""
Args:
ref mode:
b_gt (torch.Tensor): [b, c, h, w]
b_pred (torch.Tensor): [b, c, h, w]
unref mode:
sr (torch.Tensor): [b, c, h, w]
ms (torch.Tensor, optional): [b, c, h/ratio, w/ratio]
lms (torch.Tensor): [b, c, h, w]
pan (torch.Tensor): [b, c, h, w]
"""
kwargs = self._call_check_args_to_kwargs(*args)
n = args[0].shape[0]
self.acc_ave = self._sum_acc(self.acc_ave, self.once_batch_call(**kwargs), self._call_n, n2=n)
self.acc_ave = self._average_acc(self.acc_ave, self._call_n + n)
self._call_n += n
return self.acc_ave
def clear_history(self, verbose=False):
if verbose:
print('>> AccAnalysis: clear history')
self._acc_d = {}
self._call_n = 0
self.acc_ave = {'SAM': 0., 'ERGAS': 0., 'PSNR': 0., 'CC': 0., 'SSIM': 0.} if self.ref else \
{'D_S': 1., 'D_lambda': 1., 'HQNR': 0.}
def print_str(self, decimals=6):
return dict_to_str(self.acc_ave, decimals=decimals)
def __repr__(self) -> str:
repr_str = f'AnalysisPanAcc(ratio={self.ratio}, ref={self.ref}):'
repr_str += f'\n{self.print_str()}'
return repr_str
def normalize_to_01(x):
# normalize tensor to [0, 1]
if isinstance(x, torch.Tensor):
x -= x.flatten(-2).min(-1, keepdim=True)[0][..., None]
x /= x.flatten(-2).max(-1, keepdim=True)[0][..., None]
elif isinstance(x, np.ndarray):
x -= x.min((-2, -1), keepdims=True)
x /= x.max((-2, -1), keepdims=True)
return x
def psnr_one_img(img_gt, img_test):
"""
calculate PSNR for one image
:param img_gt: ground truth image, numpy array, shape [H, W, C]
:param img_test: test or inference image, numpy array, shape [H, W, C]
:return: PSNR, float type
"""
assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size'
return peak_signal_noise_ratio(img_gt, img_test)
def psnr_batch_tensor_metric(b_gt, b_pred):
"""
calculate PSNR for batch tensor images
:param b_gt: tensor, shape [B, C, H, W]
:param b_test: tensor, shape [B, C, H, W]
:return:
"""
assert b_gt.shape[0] == b_pred.shape[0]
bs = b_gt.shape[0]
psnr = 0.
for gt, t in zip(b_gt, b_pred):
psnr += psnr_one_img(*(to_numpy(gt, t)))
return psnr / bs
def ssim_one_image(img_gt, img_test, channel_axis=0):
assert img_gt.shape == img_test.shape, 'image 1 and image 2 should have the same size'
return structural_similarity(img_gt, img_test, channel_axis=channel_axis,
data_range=1.)
def ssim_batch_tensor_metric(b_gt, b_pred):
assert b_gt.shape[0] == b_pred.shape[0]
bs = b_gt.shape[0]
ssim = 0.
for gt, t in zip(b_gt, b_pred):
ssim += ssim_one_image(*(to_numpy(gt, t)), channel_axis=0)
return ssim / bs
if __name__ == '__main__':
sr = torch.rand(4, 3, 256, 256)
ms = torch.rand(4, 3, 64, 64)
lms = torch.rand(4, 3, 256, 256)
pan = torch.rand(4, 3, 256, 256)
gt = torch.rand(4, 3, 256, 256)
analysis = AnalysisPanAcc(ref=False, ratio=4, default_max_value=2047)
for i in range(2):
analysis(sr[i:i+2], ms[i:i+2], lms[i:i+2], pan[i:i+2])
print(analysis.print_str())
|
2977094657/DidaAPI | 1,671 | frontend/docs/api/auth/index.md | # 认证概览
滴答清单API提供两种认证方式,您可以根据使用场景选择合适的登录方法。
## 认证方式对比
| 特性 | 密码登录 | 微信登录 |
|------|----------|----------|
| **便利性** | 需要输入用户名密码 | 扫码即可 |
| **安全性** | 依赖密码强度 | 依赖微信安全体系 |
| **适用场景** | 自动化脚本、批量操作 | 交互式使用 |
| **实现复杂度** | 简单直接 | 复杂(需要轮询) |
| **网络要求** | 一次HTTP请求 | 多次请求和轮询 |
| **用户体验** | 传统表单登录 | 现代扫码体验 |
## 认证流程
### 密码登录流程
```mermaid
sequenceDiagram
participant User as 用户
participant Client as 客户端
participant API as API服务器
participant Dida as 滴答清单服务器
User->>Client: 1. 输入用户名密码
Client->>API: 2. POST https://api.dida365.com/api/v2/user/signon
API->>Dida: 3. 验证登录凭据
Dida-->>API: 4. 返回认证令牌
API->>API: 5. 保存会话信息
API-->>Client: 6. 返回登录结果
Client->>User: 7. 显示登录状态
```
### 微信登录流程
```mermaid
sequenceDiagram
participant User as 用户
participant Client as 客户端
participant API as API服务器
participant WeChat as 微信服务器
participant Dida as 滴答清单服务器
User->>Client: 1. 请求微信登录
Client->>API: 2. GET https://open.weixin.qq.com/connect/qrconnect
API->>WeChat: 3. 获取二维码
WeChat-->>API: 4. 返回二维码
API-->>Client: 5. 返回二维码URL
Client->>User: 6. 显示二维码
User->>WeChat: 7. 微信扫码确认
Client->>API: 8. GET https://lp.open.weixin.qq.com/connect/l/qrconnect
API->>WeChat: 9. 轮询登录状态
WeChat-->>API: 10. 返回授权码
API->>Dida: 11. 验证授权码
Dida-->>API: 12. 返回认证令牌
API->>API: 13. 保存会话信息
API-->>Client: 14. 返回登录结果
```
## 详细文档
### 密码登录相关
- [密码登录](./password-login.md) - 详细的密码登录接口文档
### 微信登录相关
- [微信登录流程](./wechat-login-flow.md) - 完整的微信扫码登录流程
- [获取微信二维码](./get-wechat-qrcode.md) - 获取微信登录二维码
- [轮询登录状态](./poll-login-status.md) - 检查二维码扫码状态
- [验证微信登录](./validate-wechat-login.md) - 验证微信登录并获取令牌
|
294coder/Efficient-MIF | 20,170 | utils/misc.py | import argparse
from contextlib import contextmanager
import json
from multiprocessing import context
import os
import os.path as osp
import random
import time
from typing import Dict, Iterable, Sequence, Union
import importlib
import h5py
import kornia.augmentation as K
from fvcore.nn import FlopCountAnalysis, flop_count_table
import yaml
import numpy as np
import torch
import torch.distributed as dist
import kornia
import shortuuid
from matplotlib import pyplot as plt
from torch.backends import cudnn
def exists(val):
return val is not None
def default(val, d):
return val if exists(val) else d
def is_none(val):
return val in ('none', 'None', 'NONE', None)
def set_all_seed(seed=2022):
torch.manual_seed(seed)
torch.cuda.manual_seed(seed)
torch.cuda.manual_seed_all(seed)
random.seed(seed)
np.random.seed(seed)
cudnn.deterministic = True
cudnn.benchmark = False
def to_numpy(*args):
l = []
for i in args:
if isinstance(i, torch.Tensor):
l.append(i.detach().cpu().numpy())
return l
def to_tensor(*args, device, dtype):
out = []
for a in args:
out.append(torch.tensor(a, dtype=dtype).to(device))
return out
def args_no_str_none(value: str) -> "str | None":
if value.lower() == "none":
return None
return value
def to_device(*args, device):
out = []
for a in args:
out.append(a.to(device))
return out
def rgb_to_ycbcr(image: torch.Tensor) -> torch.Tensor:
"""
Convert an RGB image to YCbCr.
Args:
image: RGB image tensor with shape (..., 3, H, W) in range [0, 1]
Returns:
YCbCr image tensor with shape (..., 3, H, W)
"""
if not isinstance(image, torch.Tensor):
raise TypeError(f"Input type is not a torch.Tensor. Got {type(image)}")
if len(image.shape) < 3 or image.shape[-3] != 3:
raise ValueError(f"Input size must have a shape of (..., 3, H, W). Got {image.shape}")
r: torch.Tensor = image[..., 0, :, :]
g: torch.Tensor = image[..., 1, :, :]
b: torch.Tensor = image[..., 2, :, :]
y: torch.Tensor = 0.29900 * r + 0.58700 * g + 0.11400 * b
cb: torch.Tensor = -0.168736 * r - 0.331264 * g + 0.50000 * b + 0.5
cr: torch.Tensor = 0.50000 * r - 0.418688 * g - 0.081312 * b + 0.5
return torch.stack([y, cb, cr], dim=-3)
def ycbcr_to_rgb(image: torch.Tensor) -> torch.Tensor:
"""
Convert a YCbCr image to RGB.
Args:
image: YCbCr image tensor with shape (..., 3, H, W)
Returns:
RGB image tensor with shape (..., 3, H, W) in range [0, 1]
"""
if not isinstance(image, torch.Tensor):
raise TypeError(f"Input type is not a torch.Tensor. Got {type(image)}")
if len(image.shape) < 3 or image.shape[-3] != 3:
raise ValueError(f"Input size must have a shape of (..., 3, H, W). Got {image.shape}")
y: torch.Tensor = image[..., 0, :, :]
cb: torch.Tensor = image[..., 1, :, :]
cr: torch.Tensor = image[..., 2, :, :]
r: torch.Tensor = y + 1.40200 * (cr - 0.5)
g: torch.Tensor = y - 0.34414 * (cb - 0.5) - 0.71414 * (cr - 0.5)
b: torch.Tensor = y + 1.77200 * (cb - 0.5)
return torch.stack([r, g, b], dim=-3).clamp(0, 1)
@contextmanager
def y_pred_model_colored(vis: torch.Tensor, enable: bool=True):
"""
Context manager to handle YCbCr color space conversion for image processing.
Args:
vis (torch.Tensor): Input RGB image tensor of shape (B, 3, H, W)
Yields:
torch.Tensor: Y channel of the image
The context manager handles:
1. Converting RGB to YCbCr
2. Extracting Y channel for processing
3. Converting processed Y channel back to RGB
"""
if enable:
assert vis.size(1) == 3, 'vis should be a 3-channel rgb image'
y_cb_cr = kornia.color.rgb_to_ycbcr(vis)
cbcr = y_cb_cr[:, 1:]
y = y_cb_cr[:, :1]
def back_to_rgb(pred_y):
y_cb_cr = torch.cat([pred_y, cbcr], dim=1)
return kornia.color.ycbcr_to_rgb(y_cb_cr)
else:
y = vis
def back_to_rgb(pred_rgb):
return pred_rgb
try:
# Yield the Y channel for processing
yield y, back_to_rgb
finally:
pass
class WindowBasedPadder(object):
def __init__(self, window_size=64) -> None:
self.window_size = window_size
self.padding_fn = None
def find_least_pad(self, base_size: tuple, window_size: int):
least_size = []
for b_s in base_size:
if b_s % window_size == 0:
least_size.append(b_s)
else:
mult = b_s // window_size
mult += 1
least_size.append(mult * window_size)
return least_size
def __call__(self, img: torch.Tensor, size: Sequence[int]=None, no_check_pad: bool = False):
if no_check_pad:
assert self.padding_fn is not None
return self.padding_fn(img)
if size is not None:
self._last_img_ori_size = size
self.padding_fn = K.PadTo(size)
else:
pad_size = self.find_least_pad(img.shape[-2:], self.window_size)
self._last_img_ori_size = img.shape[-2:]
self.padding_fn = K.PadTo(pad_size)
return self.padding_fn(img)
def inverse(self, img: torch.Tensor):
return self.padding_fn.inverse(img, size=self._last_img_ori_size)
def h5py_to_dict(file: h5py.File, keys=None) -> dict[str, np.ndarray]:
"""get all content in a h5py file into a dict contains key and values
Args:
file (h5py.File): h5py file
keys (list, optional): h5py file keys used to extract values.
Defaults to ["ms", "lms", "pan", "gt"].
Returns:
dict[str, np.ndarray]:
"""
d = {}
if keys is None:
keys = list(file.keys())
for k in keys:
v = file[k][:]
d[k] = v
return d
def dict_to_str(d, decimals=4):
n = len(d)
func = lambda k, v: f"{k}: {torch.round(v, decimals=decimals).item() if isinstance(v, torch.Tensor) else np.round(v, decimals=decimals)}"
s = ""
for i, (k, v) in enumerate(d.items()):
s += func(k, v) + (", " if i < n - 1 else "")
return s
def prefixed_dict_key(d, prefix, sep="_"):
# e.g.
# SSIM -> train_SSIM
d2 = {}
for k, v in d.items():
d2[prefix + sep + k] = v
return d2
# TODO: nees test
class CheckPointManager(object):
def __init__(
self,
model: torch.nn.Module,
save_path: str,
save_every_eval: bool = False,
verbose: bool = True,
):
"""
manage model checkpoints
Args:
model: nn.Module, can be single node model or multi-nodes model
save_path: str like '/home/model_ckpt/resnet.pth' or '/home/model_ckpt/exp1' when @save_every_eval
is False or True
save_every_eval: when False, save params only when ep_loss is less than optim_loss.
when True, save params every eval epoch
verbose: print out all information
e.g.
@save_every_eval=False, @save_path='/home/ckpt/resnet.pth'
weights will be saved like
-------------
/home/ckpt
|-resnet.pth
-------------
@save_every_eval=True, @save_path='/home/ckpt/resnet'
weights will be saved like
-------------
/home/ckpt
|-resnet
|-ep_20.pth
|-ep_40.pth
-------------
"""
self.model = model
self.save_path = save_path
self.save_every_eval = save_every_eval
self._optim_loss = torch.inf
self.verbose = verbose
self.check_path_legal()
def check_path_legal(self):
if self.save_every_eval:
if not os.path.exists(self.save_path):
os.makedirs(self.save_path)
else:
assert self.save_path.endswith(".pth")
par_dir = os.path.dirname(self.save_path)
if not os.path.exists(par_dir):
os.makedirs(par_dir)
def save(
self,
ep_loss: Union[float, torch.Tensor] = None,
ep: int = None,
extra_saved_dict: dict = None,
):
"""
Args:
ep_loss: should be set when @save_every_eval=False
ep: should be set when @save_every_eval=True
extra_saved_dict: a dict which contains other information you want to save with model
e.g. {'optimizer_ckpt': op_ckpt, 'time': '2023/1/21'}
Returns:
"""
if isinstance(ep_loss, torch.Tensor):
ep_loss = ep_loss.item()
saved_dict = {}
if not self.save_every_eval:
assert ep_loss is not None
if ep_loss < self._optim_loss:
self._optim_loss = ep_loss
path = self.save_path
saved_dict["optim_loss"] = ep_loss
else:
print(
"optim loss: {}, now loss: {}, not saved".format(
self._optim_loss, ep_loss
)
)
return
else:
assert ep is not None
path = os.path.join(self.save_path, "ep_{}.pth".format(ep))
if extra_saved_dict is not None:
assert "model" not in list(saved_dict.keys())
saved_dict = extra_saved_dict
try:
saved_dict["model"] = self.model.module.state_dict()
except:
saved_dict["model"] = self.model.state_dict()
torch.save(saved_dict, path)
if self.verbose:
print(
f"saved params contains\n",
*[
"\t -{}: {}\n".format(k, v if k != "model" else "model params")
for k, v in saved_dict.items()
],
"saved path: {}".format(path),
)
def is_main_process(func=None):
"""
check if current process is main process in ddp
warning: if not in ddp mode, always return True
:return:
"""
def _is_main_proc():
if dist.is_initialized():
return dist.get_rank() == 0
else:
return True
if func is None:
return _is_main_proc()
else:
def warp_func(*args, **kwargs):
if _is_main_proc():
return func(*args, **kwargs)
else:
return None
return warp_func
def print_args(args):
d = args.__dict__
for k, v in d.items():
print(f"{k}: {v}")
def yaml_load(name, base_path="./configs", end_with="_config.yaml"):
path = osp.join(base_path, name + end_with)
if osp.exists(path):
f = open(path)
cont = f.read()
return yaml.load(cont, Loader=yaml.FullLoader)
else:
print("configuration file not exists")
raise FileNotFoundError(f'file not exists: {path}')
def json_load(name, base_path="./configs"):
path = osp.join(base_path, name + "_config.json")
with open(path) as f:
return json.load(f)
def config_py_load(name, base_path="configs"):
args = importlib.import_module(f".{name}_config", package=base_path)
return args.config
class NameSpace:
def __init__(self, **kwargs):
self.__dict__.update(kwargs)
@property
def attrs(self):
return self.__dict__
def to_dict(self):
out = {}
d = self.attrs
for k, v in d.items():
if isinstance(v, NameSpace):
out[k] = v.to_dict()
else:
out[k] = v
return out
def __repr__(self, d=None, nprefix=0):
repr_str = ""
if d is None:
d = self.attrs
for k, v in d.items():
if isinstance(v, NameSpace):
repr_str += (
" " * nprefix
+ f"{k}: \n"
+ f"{self.__repr__(v.attrs, nprefix + 1)}"
)
else:
repr_str += " " * nprefix + f"{k}: {v}\n"
return repr_str
def __getitem__(self, item):
return self.attrs[item]
def __setitem__(self, key, value):
setattr(self.attrs, key, value)
def recursive_search_dict2namespace(d: Dict):
"""
convert a yaml-like configuration (dict) to namespace-like class
e.g.
{'lr': 1e-3, 'path': './datasets/train_wv3.h5'} ->
NameSpace().lr = 1e-3, NameSpace().path = './datasets/train_wv3.h5'
Warning: the value in yaml-like configuration should not be another dict
:param d:
:return:
"""
namespace = NameSpace()
for k, v in d.items():
if isinstance(v, dict):
setattr(namespace, k, recursive_search_dict2namespace(v))
else:
setattr(namespace, k, v)
return namespace
def merge_args_namespace(parser_args: argparse.Namespace, namespace_args: NameSpace):
"""
merge parser_args and self-made class _NameSpace configurations together for better
usage.
return args that support dot its member, like args.optimizer.lr
:param parser_args:
:param namespace_args:
:return:
"""
# namespace_args.__dict__.update(parser_args.__dict__)
namespace_d = namespace_args.__dict__
for k, v in parser_args.__dict__.items():
if not (k in namespace_d.keys() and v is None):
setattr(namespace_args, k, v)
return namespace_args
def generate_id(length: int = 8) -> str:
# ~3t run ids (36**8)
run_gen = shortuuid.ShortUUID(alphabet=list("0123456789abcdefghijklmnopqrstuvwxyz"))
return str(run_gen.random(length))
def find_weight(weight_dir="./weight/", id=None, func=None):
"""
return weight absolute path referring to id
Args:
weight_dir: weight dir that saved weights
id: weight id
func: split string function
Returns: str, absolute path
"""
assert id is not None, "@id can not be None"
weight_list = os.listdir(weight_dir)
if func is None:
func = lambda x: x.split(".")[0].split("_")[-1]
for id_s in weight_list:
only_id = func(id_s)
if only_id == id:
return os.path.abspath(os.path.join(weight_dir, id_s))
print(f"can not find {id}")
return None
def _delete_unneeded_weight_file(weight_dir="./weight/", id=None):
"""
delete unneeded weight file referring to id
Args:
weight_dir:
id:
Returns:
"""
assert id is not None, "@id can not be None"
abspath = find_weight(weight_dir, id)
if abspath is not None:
assert os.path.exists(abspath)
os.remove(abspath)
print(f"delete {os.path.basename(abspath)}")
def print_network_params_macs_fvcore(network, *inputs):
"""
print out network's parameters and macs by using
fvcore package
Args:
network: nn.Module
*inputs: input argument
Returns: None
"""
analysis = FlopCountAnalysis(network, inputs=inputs)
print(flop_count_table(analysis))
def clip_dataset_into_small_patches(
file: h5py.File,
patch_size: int,
up_ratio: int,
ms_channel: int,
pan_channel: int,
dataset_keys: Union[list[str], tuple[str]] = ("gt", "ms", "lms", "pan"),
save_path: str = "./data/clip_data.h5",
):
"""
clip patches at spatial dim
Args:
file: h5py.File of original dataset
patch_size: ms clipped size
up_ratio: shape of lms divide shape of ms
ms_channel:
pan_channel:
dataset_keys: similar to [gt, ms, lms, pan]
save_path: must end with h5
Returns:
"""
unfold_fn = lambda x, c, ratio: (
torch.nn.functional.unfold(
x, kernel_size=patch_size * ratio, stride=patch_size * ratio
)
.transpose(1, 2)
.reshape(-1, c, patch_size * ratio, patch_size * ratio)
)
assert len(dataset_keys) == 4, "length of @dataset_keys should be 4"
assert save_path.endswith("h5"), "saved file should end with h5 but get {}".format(
save_path.split(".")[-1]
)
gt = unfold_fn(torch.tensor(file[dataset_keys[0]][:]), ms_channel, up_ratio)
ms = unfold_fn(torch.tensor(file[dataset_keys[1]][:]), ms_channel, 1)
lms = unfold_fn(torch.tensor(file[dataset_keys[2]][:]), ms_channel, up_ratio)
pan = unfold_fn(torch.tensor(file[dataset_keys[3]][:]), pan_channel, up_ratio)
print("clipped datasets shape:")
print("{:^20}{:^20}{:^20}{:^20}".format(*[k for k in dataset_keys]))
print(
"{:^20}{:^20}{:^20}{:^20}".format(
str(gt.shape), str(ms.shape), str(lms.shape), str(pan.shape)
)
)
base_path = os.path.dirname(save_path)
if not os.path.exists(base_path):
os.makedirs(base_path)
print(f"make dir {base_path}")
save_file = h5py.File(save_path, "w")
for k, data in zip(dataset_keys, [gt, ms, lms, pan]):
save_file.create_dataset(name=k, data=data)
print(f"create data {k}")
file.close()
save_file.close()
print("file closed")
def dist_gather_object(obj, n_ranks=1, dest=0, all_gather=False):
def _iter_tensor_to_rank(rank_obj, dest=0):
if isinstance(rank_obj, dict):
for k, v in rank_obj.items():
if isinstance(v, torch.Tensor):
rank_obj[k] = v.to(dest)
elif isinstance(v, Iterable):
rank_obj[k] = _iter_tensor_to_rank(v, dest)
elif isinstance(rank_obj, (list, tuple)):
if isinstance(rank_obj, tuple):
rank_obj = list(rank_obj)
for i, v in enumerate(rank_obj):
if isinstance(v, torch.Tensor):
rank_obj[i] = v.to(dest)
elif isinstance(v, Iterable):
rank_obj[i] = _iter_tensor_to_rank(v, dest)
elif isinstance(rank_obj, torch.Tensor):
rank_obj = rank_obj.to(dest)
return rank_obj
if n_ranks == 1:
return obj
elif n_ranks > 1:
rank_objs = [None] * n_ranks
if all_gather:
# all proc to proc dest
dist.all_gather_object(rank_objs, obj)
# if is_main_process():
# _scattered_objs_lst = [rank_objs] * n_ranks
# else:
# _scattered_objs_lst = [None] * n_ranks
# received_objs = [None]
# dist.scatter_object_list(received_objs, _scattered_objs_lst)
rank_objs = _iter_tensor_to_rank(rank_objs, dest=dest)
else:
dist.gather_object(obj, rank_objs if is_main_process() else None, dest)
if is_main_process():
rank_objs = _iter_tensor_to_rank(rank_objs, dest)
return rank_objs
else:
raise ValueError("n_ranks should be greater than 0")
if __name__ == "__main__":
# path = "/home/ZiHanCao/datasets/HISI/new_harvard/x8/test_harvard(with_up)x8_rgb.h5"
# file = h5py.File(path)
# clip_dataset_into_small_patches(
# file,
# patch_size=16,
# up_ratio=8,
# ms_channel=31,
# pan_channel=3,
# dataset_keys=["GT", "LRHSI", "HSI_up", "RGB"],
# save_path="/home/ZiHanCao/datasets/HISI/new_harvard/x8/test_clip_128.h5",
# )
# vis = torch.randn(1, 3, 256, 256).clip(0, 1)
# ir = torch.randn(1, 1, 256, 256).clip(0, 1)
# model = lambda vis, ir: vis
# with y_pred_model_colored(vis, enable=True) as (y, back_to_rgb):
# pred_y = model(y, ir)
# pred_rgb = back_to_rgb(pred_y)
# # assert equal
# print(torch.isclose(pred_rgb, vis))
# mean_diff = torch.mean(torch.abs(vis - pred_rgb))
# print("mean difference:", mean_diff.item())
d = dict(
a=1, b=2,
c=dict(
ca=1,
cb=2,
)
)
args = NameSpace(**d)
print(args.a)
print(args['c']['ca']) |
294coder/Efficient-MIF | 3,738 | utils/calculate_metric_from_file.py | #%%
import numpy as np
import torch
import matplotlib.pyplot as plt
from scipy.io import loadmat
from pathlib import Path
import h5py
from tabulate import tabulate
from tqdm import tqdm, trange
from utils import AnalysisPanAcc, find_data_path, h5py_to_dict
def metric_dicts_ave_and_std(metrics: list[dict]):
ave = {}
std = {}
keys = metrics[0].keys()
for k in keys:
ave[k] = np.mean([m[k] for m in metrics])
std[k] = np.std([m[k] for m in metrics])
return ave, std
def to_tabulate(ave: dict, std: dict):
table = []
for k in ave.keys():
table.append([k, ave[k], std[k]])
return tabulate(table, headers=["Metric", "Average", "Standard Deviation"], tablefmt="rounded_grid")
def norm_to_0_1(*args, norm_const=2047):
return [a / norm_const for a in args] if len(args) > 1 else args[0] / norm_const
#%%
path = '/volsparse1/czh/exps/fcformer-bk/visualized_img/data_MIMO_SST_harvard_x8_ref.mat'
full_res = True if 'unref' in path else False
dataset_type = 'harvard_x8'
ratio = 4 if dataset_type not in ['cave', 'harvard'] else int(dataset_type.split('_')[-1][-1])
print('===================DATASET========================')
print(f'Dataset: {dataset_type}')
print(f'Full Resolution: {full_res}')
print(f'Ratio: {ratio}')
print('=================================================')
const = {'wv3': 2047, 'gf': 1023, 'qb': 1023, 'wv2': 2047,
'cave': 1, 'harvard': 1, 'gf5': 1}.get(dataset_type.split('_')[0], 1)
dataset_path = find_data_path(dataset_type, full_res)
dataset = h5py.File(dataset_path, 'r')
if dataset_type[:4] == "cave" or dataset_type[:7] == "harvard":
keys = ["LRHSI", "HSI_up", "RGB", "GT"]
else:
keys = None
dataset = h5py_to_dict(dataset, keys)
analysis = AnalysisPanAcc(ref=not full_res, ratio=ratio, sensor=dataset_type)
path = Path(path)
if path.is_dir():
file_lst = list(path.glob('*.mat'))
metrics = []
for i, path in tqdm(enumerate(file_lst)):
if hasattr(dataset, 'ms'):
ms, lms, pan = dataset["ms"][i:i+1], dataset["lms"][i:i+1], dataset["pan"][i:i+1]
else:
ms, lms, pan = dataset["LRHSI"][i:i+1], dataset["HSI_up"][i:i+1], dataset["RGB"][i:i+1]
sr = loadmat(path)['sr']
ms, lms, pan, sr = norm_to_0_1(ms, lms, pan, sr, norm_const=const)
if full_res:
analysis(sr, ms, lms, pan)
else:
if hasattr(dataset, 'gt'):
gt = dataset["gt"][i:i+1]
else:
gt = dataset['GT'][i:i+1]
gt = norm_to_0_1(gt, norm_const=const)
analysis(sr, gt)
metrics.append(analysis.acc_ave)
analysis.clear_history()
ave, std = metric_dicts_ave_and_std(metrics)
table = to_tabulate(ave, std)
else:
files = loadmat(path)['sr']
metrics = []
for i in trange(len(files)):
if hasattr(dataset, 'ms'):
ms, lms, pan = dataset["ms"][i:i+1], dataset["lms"][i:i+1], dataset["pan"][i:i+1]
else:
ms, lms, pan = dataset["LRHSI"][i:i+1], dataset["HSI_up"][i:i+1], dataset["RGB"][i:i+1]
sr = files[i:i+1]
ms, lms, pan, sr = norm_to_0_1(ms, lms, pan, sr, norm_const=const)
if full_res:
analysis(sr, ms, lms, pan)
else:
if hasattr(dataset, 'gt'):
gt = dataset["gt"][i:i+1]
else:
gt = dataset['GT'][i:i+1]
gt = norm_to_0_1(gt, norm_const=const)
analysis(sr, gt)
metrics.append(analysis.acc_ave)
analysis.clear_history()
ave, std = metric_dicts_ave_and_std(metrics)
table = to_tabulate(ave, std)
print(table)
|
2977094657/DidaAPI | 2,651 | frontend/docs/api/auth/validate-wechat-login.md | # 验证微信登录
使用从轮询接口获得的授权码向滴答清单验证微信登录,获取认证令牌。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/user/sign/wechat/validate`
- **请求方法**: `GET`
- **认证要求**: 无需认证(此接口用于获取认证)
- **所属平台**: 滴答清单
## 请求参数
### Query Parameters
| 参数名 | 类型 | 必填 | 说明 |
|--------|------|------|------|
| code | string | 是 | 微信授权码,从轮询接口获得 |
| state | string | 否 | 状态参数,需与获取二维码时一致 |
## 完整请求示例
```http
GET https://api.dida365.com/api/v2/user/sign/wechat/validate?code={authorization_code}&state={state} HTTP/1.1
Host: api.dida365.com
Accept: */*
Accept-Encoding: gzip, deflate, br, zstd
Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7
Content-Type: application/json
Origin: https://dida365.com
Referer: https://dida365.com/
User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36
X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":""}
```
## 响应格式
### 成功响应
**状态码**: `200 OK`
**Content-Type**: `application/json;charset=UTF-8`
### 响应头(重要)
响应头中包含认证相关的Cookie:
```http
Set-Cookie: t={auth_token}; Domain=.dida365.com; Expires={expires_date}; Path=/; Secure; HttpOnly; SameSite=None
Set-Cookie: _csrf_token={csrf_token}; Domain=.dida365.com; Expires={expires_date}; Path=/; Secure
```
### 响应体
```json
{
"success": true,
"message": "string",
"user": {
"id": "string",
"username": "string",
"email": "string"
}
}
```
## 关键信息提取
### 认证令牌 (t cookie)
这是最重要的认证信息,用于后续所有API调用:
```
t={auth_token}
```
### CSRF令牌 (_csrf_token cookie)
用于防止跨站请求伪造攻击:
```
_csrf_token={csrf_token}
```
## Cookie解析
### 解析Set-Cookie头
```python
import re
def parse_cookies(set_cookie_header):
"""解析Set-Cookie头"""
cookies = {}
# 分割多个cookie
cookie_parts = set_cookie_header.split(', ')
for part in cookie_parts:
# 提取cookie名称和值
match = re.match(r'([^=]+)=([^;]+)', part)
if match:
name = match.group(1).strip()
value = match.group(2).strip()
# 处理空值cookie(用于删除)
if value != '""':
cookies[name] = value
return cookies
```
## 后续API调用
获得认证令牌后,所有需要认证的API调用都需要包含:
### 必需的Cookie
```http
Cookie: t={认证令牌}; _csrf_token={CSRF令牌}
```
### 必需的请求头
```http
X-Csrftoken: {CSRF令牌}
X-Device: {"platform":"web","os":"Windows 10","device":"Chrome 136.0.0.0","name":"","version":6310,"id":"{device_id}","channel":"website","campaign":"","websocket":""}
```
## 相关接口
- [轮询登录状态](./poll-login-status.md) - 获取用于验证的授权码
- [获取所有任务](../tasks/get-all-tasks.md) - 使用认证令牌获取任务数据
|
2977094657/DidaAPI | 1,427 | frontend/docs/api/auth/wechat-callback.md | # 微信登录回调处理
## 接口信息
- **接口路径**: `GET /auth/wechat/callback`
- **接口描述**: 处理微信扫码后的回调,提取code参数
- **请求方式**: GET
- **认证要求**: 无需认证
## 请求参数
### Query参数
| 参数名 | 类型 | 必填 | 描述 | 示例值 |
|--------|------|------|------|--------|
| code | string | 是 | 微信返回的授权码 | `001uZIkl2urXGf4qrmml2mAIWz4uZIkW` |
| state | string | 否 | 状态参数 | `Lw==` |
## 请求示例
```bash
curl -X GET "http://localhost:8000/auth/wechat/callback?code=001uZIkl2urXGf4qrmml2mAIWz4uZIkW&state=Lw=="
```
## 响应格式
### 成功响应
```json
{
"message": "微信登录成功",
"code": "001uZIkl2urXGf4qrmml2mAIWz4uZIkW",
"state": "Lw==",
"login_result": {
"success": true,
"message": "微信登录验证成功",
"user_info": {
"username": "用户名",
"email": "user@example.com"
},
"session_info": {
"auth_token": "43A001113F9610FFC85AA97B18A297A4...",
"csrf_token": "rMwhKGWevLOhHIhFv6hHjAziGDbkpnuY..."
}
},
"next_step": "用户已成功登录,可以访问受保护的资源"
}
```
### 失败响应
```json
{
"message": "微信登录失败",
"code": "001uZIkl2urXGf4qrmml2mAIWz4uZIkW",
"state": "Lw==",
"error": "验证失败"
}
```
## 使用说明
1. **回调处理**: 这个接口用于演示微信回调的处理流程
2. **自动验证**: 接收到回调后会自动进行登录验证
3. **会话建立**: 验证成功后会建立用户会话
4. **实际使用**: 在实际使用中,微信会重定向到配置的redirect_uri
## 相关接口
- [微信登录流程](./wechat-login-flow.md)
- [获取微信二维码](./get-wechat-qrcode.md)
- [轮询登录状态](./poll-login-status.md)
- [验证微信登录](./validate-wechat-login.md)
## 注意事项
- 这是微信OAuth流程的最后一步
- code参数只能使用一次,有效期很短
- 成功后会自动建立认证会话
- 可以直接用于后续API调用
|
2977094657/DidaAPI | 1,931 | frontend/docs/api/auth/poll-login-status.md | # 轮询微信登录状态
通过长轮询检查用户是否已扫码并确认微信登录。
## 接口信息
- **接口URL**: `https://long.open.weixin.qq.com/connect/l/qrconnect`
- **请求方法**: `GET`
- **认证要求**: 无需认证
- **所属平台**: 微信开放平台
## 请求参数
### Query Parameters
| 参数名 | 类型 | 必填 | 说明 |
|--------|------|------|------|
| uuid | string | 是 | 二维码密钥,从获取二维码接口中提取的16位字符串 |
| _ | number | 是 | 时间戳,用于防止缓存,格式:毫秒级时间戳 |
## 完整请求示例
```http
GET https://long.open.weixin.qq.com/connect/l/qrconnect?uuid={qr_code_key}&_={timestamp} HTTP/1.1
Host: long.open.weixin.qq.com
User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36
Accept: */*
Accept-Language: zh-CN,zh;q=0.9,en;q=0.8
Referer: https://open.weixin.qq.com/
```
## 响应格式
### 响应类型
**Content-Type**: `text/javascript; charset=utf-8`
**响应体**: JavaScript代码片段
### 状态码说明
响应中包含 `window.wx_errcode` 表示不同的状态:
| errcode | 说明 | 后续操作 |
|---------|------|----------|
| 404 | 等待扫码 | 继续轮询 |
| 403 | 已扫码,等待用户确认 | 继续轮询 |
| 405 | 登录成功,获得授权码 | 停止轮询,提取授权码 |
| 408 | 二维码已过期 | 停止轮询,重新获取二维码 |
| 400 | 二维码已失效 | 停止轮询,重新获取二维码 |
## 响应示例
### 等待扫码 (errcode=404)
```javascript
window.wx_errcode=404;window.wx_code='';
```
### 已扫码等待确认 (errcode=403)
```javascript
window.wx_errcode=403;window.wx_code='';
```
### 登录成功 (errcode=405)
```javascript
window.wx_errcode=405;window.wx_code='{authorization_code}';
```
### 二维码过期 (errcode=408)
```javascript
window.wx_errcode=408;window.wx_code='';
```
## 响应解析
### 提取状态码和授权码
```javascript
// 解析响应文本
function parseWeChatResponse(responseText) {
const errcodeMatch = responseText.match(/window\.wx_errcode\s*=\s*(\d+)/);
const codeMatch = responseText.match(/window\.wx_code\s*=\s*'([^']*)'/);
return {
errcode: errcodeMatch ? parseInt(errcodeMatch[1]) : null,
code: codeMatch ? codeMatch[1] : ''
};
}
```
## 相关接口
- [获取微信二维码](./get-wechat-qrcode.md) - 获取用于轮询的二维码密钥
- [验证微信登录](./validate-wechat-login.md) - 使用获得的授权码进行登录验证
|
2977094657/DidaAPI | 2,177 | frontend/docs/api/auth/get-wechat-qrcode.md | # 获取微信登录二维码
获取微信扫码登录的二维码,这是微信登录流程的第一步。
## 接口信息
- **接口URL**: `https://open.weixin.qq.com/connect/qrconnect`
- **请求方法**: `GET`
- **认证要求**: 无需认证
- **所属平台**: 微信开放平台
## 请求参数
### Query Parameters
| 参数名 | 类型 | 必填 | 说明 |
|--------|------|------|------|
| appid | string | 是 | 微信应用ID |
| redirect_uri | string | 是 | 授权后重定向的回调地址 |
| response_type | string | 是 | 返回类型,固定值:`code` |
| scope | string | 是 | 应用授权作用域,固定值:`snsapi_login` |
| state | string | 否 | 重定向后会带上state参数,开发者可以填写a-zA-Z0-9的参数值,最多128字节 |
## 完整请求示例
```http
GET https://open.weixin.qq.com/connect/qrconnect?appid={appid}&redirect_uri={redirect_uri}&response_type=code&scope=snsapi_login&state={state} HTTP/1.1
Host: open.weixin.qq.com
User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/136.0.0.0 Safari/537.36
Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8
Accept-Language: zh-CN,zh;q=0.9,en;q=0.8,zh-TW;q=0.7
Accept-Encoding: gzip, deflate, br, zstd
```
## 响应格式
### 成功响应
**状态码**: `200 OK`
**Content-Type**: `text/html; charset=utf-8`
**响应体**: HTML页面,包含二维码图片
### 关键HTML结构
```html
<!DOCTYPE html>
<html>
<head>
<title>微信登录</title>
</head>
<body>
<!-- 其他HTML内容 -->
<img class="qrcode lightBorder js_qrcode_img" src="/connect/qrcode/{qr_code_key}">
<!-- 其他HTML内容 -->
</body>
</html>
```
## 响应解析
### 提取二维码密钥
从响应HTML中提取二维码图片的src属性:
```javascript
// 正则表达式匹配
const pattern = /<img[^>]*class="[^"]*qrcode[^"]*"[^>]*src="([^"]*)"/;
const match = html.match(pattern);
if (match) {
const srcUrl = match[1]; // "/connect/qrcode/{qr_code_key}"
const qrCodeKey = srcUrl.split('/').pop(); // "{qr_code_key}"
}
```
### 构建完整二维码URL
```javascript
const qrCodeUrl = `https://open.weixin.qq.com${srcUrl}`;
// 结果: https://open.weixin.qq.com/connect/qrcode/{qr_code_key}
```
## 响应头示例
```http
HTTP/1.1 200 OK
Content-Type: text/html; charset=utf-8
Cache-Control: no-cache, must-revalidate
Content-Encoding: gzip
Content-Length: 17543
X-Wx-Fj: 001,018,0000001024
```
## 后续步骤
获取二维码后,需要:
1. 显示二维码供用户扫描
2. 使用二维码密钥进行[轮询登录状态](./poll-login-status.md)
3. 获取授权码后进行[验证微信登录](./validate-wechat-login.md)
|
2977094657/DidaAPI | 1,914 | frontend/docs/api/auth/wechat-login-flow.md | # 微信扫码登录完整流程
微信扫码登录到滴答清单是一个多步骤的过程,涉及微信开放平台和滴答清单API的协作。
## 完整流程概览
```mermaid
sequenceDiagram
participant User as 用户
participant Client as 客户端
participant WeChat as 微信服务器
participant Dida as 滴答清单服务器
User->>Client: 1. 请求登录
Client->>WeChat: 2. 获取二维码
WeChat-->>Client: 3. 返回二维码HTML
Client->>Client: 4. 解析二维码URL
Client->>User: 5. 显示二维码
loop 轮询检查
Client->>WeChat: 6. 轮询登录状态
WeChat-->>Client: 7. 返回状态
end
User->>WeChat: 8. 微信扫码确认
WeChat-->>Client: 9. 返回授权码
Client->>Dida: 10. 验证授权码
Dida-->>Client: 11. 返回认证令牌
Client->>Client: 12. 保存认证信息
```
## 详细步骤
### 步骤1: 获取微信二维码
**接口**: [获取微信二维码](./get-wechat-qrcode.md)
请求微信开放平台获取登录二维码:
```http
GET https://open.weixin.qq.com/connect/qrconnect?appid=wxf1429a73d311aad4&redirect_uri=https://dida365.com/sign/wechat&response_type=code&scope=snsapi_login&state=Lw==
```
### 步骤2: 解析二维码信息
从返回的HTML中提取二维码图片URL:
```html
<img class="qrcode lightBorder js_qrcode_img" src="/connect/qrcode/071cfvEZ3xaPll2J">
```
提取16位密钥:`071cfvEZ3xaPll2J`
### 步骤3: 轮询登录状态
**接口**: [轮询登录状态](./poll-login-status.md)
使用长轮询检查用户是否扫码:
```http
GET https://long.open.weixin.qq.com/connect/l/qrconnect?uuid=071cfvEZ3xaPll2J&_=1748955598000
```
### 步骤4: 用户扫码确认
用户使用微信扫描二维码并确认登录。
### 步骤5: 获取授权码
轮询接口返回授权码:
```javascript
window.wx_errcode=405;window.wx_code='001uZIkl2urXGf4qrmml2mAIWz4uZIkW';
```
### 步骤6: 验证登录
**接口**: [验证微信登录](./validate-wechat-login.md)
使用授权码向滴答清单验证登录:
```http
GET https://api.dida365.com/api/v2/user/sign/wechat/validate?code=001uZIkl2urXGf4qrmml2mAIWz4uZIkW&state=Lw==
```
### 步骤7: 获取认证令牌
验证成功后获得认证令牌和CSRF令牌:
```http
Set-Cookie: t=43A001113F9610FFC85AA97B18A297A4...; Domain=.dida365.com
Set-Cookie: _csrf_token=rMwhKGWevLOhHIhFv6hHjAziGDbkpnuY...; Domain=.dida365.com
```
## 相关接口
- [获取微信二维码](./get-wechat-qrcode.md)
- [轮询登录状态](./poll-login-status.md)
- [验证微信登录](./validate-wechat-login.md)
|
281677160/openwrt-package | 94,932 | luci-app-passwall/root/usr/share/passwall/app.sh | #!/bin/sh
# Copyright (C) 2018-2020 L-WRT Team
# Copyright (C) 2021-2025 xiaorouji
. $IPKG_INSTROOT/lib/functions.sh
. $IPKG_INSTROOT/lib/functions/service.sh
CONFIG=passwall
TMP_PATH=/tmp/etc/$CONFIG
TMP_BIN_PATH=$TMP_PATH/bin
TMP_SCRIPT_FUNC_PATH=$TMP_PATH/script_func
TMP_ROUTE_PATH=$TMP_PATH/route
TMP_ACL_PATH=$TMP_PATH/acl
TMP_IFACE_PATH=$TMP_PATH/iface
TMP_PATH2=/tmp/etc/${CONFIG}_tmp
GLOBAL_ACL_PATH=${TMP_ACL_PATH}/default
LOG_FILE=/tmp/log/$CONFIG.log
APP_PATH=/usr/share/$CONFIG
RULES_PATH=/usr/share/${CONFIG}/rules
LUA_UTIL_PATH=/usr/lib/lua/luci/passwall
UTIL_SINGBOX=$LUA_UTIL_PATH/util_sing-box.lua
UTIL_SS=$LUA_UTIL_PATH/util_shadowsocks.lua
UTIL_XRAY=$LUA_UTIL_PATH/util_xray.lua
UTIL_TROJAN=$LUA_UTIL_PATH/util_trojan.lua
UTIL_NAIVE=$LUA_UTIL_PATH/util_naiveproxy.lua
UTIL_HYSTERIA2=$LUA_UTIL_PATH/util_hysteria2.lua
UTIL_TUIC=$LUA_UTIL_PATH/util_tuic.lua
echolog() {
local d="$(date "+%Y-%m-%d %H:%M:%S")"
echo -e "$d: $*" >>$LOG_FILE
}
config_get_type() {
local ret=$(uci -q get "${CONFIG}.${1}" 2>/dev/null)
echo "${ret:=$2}"
}
config_n_get() {
local ret=$(uci -q get "${CONFIG}.${1}.${2}" 2>/dev/null)
echo "${ret:=$3}"
}
config_t_get() {
local index=${4:-0}
local ret=$(uci -q get "${CONFIG}.@${1}[${index}].${2}" 2>/dev/null)
echo "${ret:=${3}}"
}
config_t_set() {
local index=${4:-0}
local ret=$(uci -q set "${CONFIG}.@${1}[${index}].${2}=${3}" 2>/dev/null)
}
get_enabled_anonymous_secs() {
uci -q show "${CONFIG}" | grep "${1}\[.*\.enabled='1'" | cut -d '.' -sf2
}
get_host_ip() {
local host=$2
local count=$3
[ -z "$count" ] && count=3
local isip=""
local ip=$host
if [ "$1" == "ipv6" ]; then
isip=$(echo $host | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}")
if [ -n "$isip" ]; then
isip=$(echo $host | cut -d '[' -f2 | cut -d ']' -f1)
fi
else
isip=$(echo $host | grep -E "([0-9]{1,3}[\.]){3}[0-9]{1,3}")
fi
[ -z "$isip" ] && {
local t=4
[ "$1" == "ipv6" ] && t=6
local vpsrip=$(resolveip -$t -t $count $host | awk 'NR==1{print}')
ip=$vpsrip
}
echo $ip
}
get_node_host_ip() {
local ip
local address=$(config_n_get $1 address)
[ -n "$address" ] && {
local use_ipv6=$(config_n_get $1 use_ipv6)
local network_type="ipv4"
[ "$use_ipv6" == "1" ] && network_type="ipv6"
ip=$(get_host_ip $network_type $address)
}
echo $ip
}
get_ip_port_from() {
local __host=${1}; shift 1
local __ipv=${1}; shift 1
local __portv=${1}; shift 1
local __ucipriority=${1}; shift 1
local val1 val2
if [ -n "${__ucipriority}" ]; then
val2=$(config_n_get ${__host} port $(echo $__host | sed -n 's/^.*[:#]\([0-9]*\)$/\1/p'))
val1=$(config_n_get ${__host} address "${__host%%${val2:+[:#]${val2}*}}")
else
val2=$(echo $__host | sed -n 's/^.*[:#]\([0-9]*\)$/\1/p')
val1="${__host%%${val2:+[:#]${val2}*}}"
fi
eval "${__ipv}=\"$val1\"; ${__portv}=\"$val2\""
}
host_from_url(){
local f=${1}
## Remove protocol part of url ##
f="${f##http://}"
f="${f##https://}"
f="${f##ftp://}"
f="${f##sftp://}"
## Remove username and/or username:password part of URL ##
f="${f##*:*@}"
f="${f##*@}"
## Remove rest of urls ##
f="${f%%/*}"
echo "${f%%:*}"
}
hosts_foreach() {
local __hosts
eval "__hosts=\$${1}"; shift 1
local __func=${1}; shift 1
local __default_port=${1}; shift 1
local __ret=1
[ -z "${__hosts}" ] && return 0
local __ip __port
for __host in $(echo $__hosts | sed 's/[ ,]/\n/g'); do
get_ip_port_from "$__host" "__ip" "__port"
eval "$__func \"${__host}\" \"\${__ip}\" \"\${__port:-${__default_port}}\" \"$@\""
__ret=$?
[ ${__ret} -ge ${ERROR_NO_CATCH:-1} ] && return ${__ret}
done
}
check_host() {
local f=${1}
a=$(echo $f | grep "\/")
[ -n "$a" ] && return 1
# 判断是否包含汉字~
local tmp=$(echo -n $f | awk '{print gensub(/[!-~]/,"","g",$0)}')
[ -n "$tmp" ] && return 1
return 0
}
get_first_dns() {
local __hosts_val=${1}; shift 1
__first() {
[ -z "${2}" ] && return 0
echo "${2}#${3}"
return 1
}
eval "hosts_foreach \"${__hosts_val}\" __first \"$@\""
}
get_last_dns() {
local __hosts_val=${1}; shift 1
local __first __last
__every() {
[ -z "${2}" ] && return 0
__last="${2}#${3}"
__first=${__first:-${__last}}
}
eval "hosts_foreach \"${__hosts_val}\" __every \"$@\""
[ "${__first}" == "${__last}" ] || echo "${__last}"
}
check_port_exists() {
local port=$1
local protocol=$2
[ -n "$protocol" ] || protocol="tcp,udp"
local result=
if [ "$protocol" = "tcp" ]; then
result=$(netstat -tln | grep -c ":$port ")
elif [ "$protocol" = "udp" ]; then
result=$(netstat -uln | grep -c ":$port ")
elif [ "$protocol" = "tcp,udp" ]; then
result=$(netstat -tuln | grep -c ":$port ")
fi
echo "${result}"
}
get_new_port() {
local port=$1
[ "$port" == "auto" ] && port=2082
local protocol=$(echo $2 | tr 'A-Z' 'a-z')
local result=$(check_port_exists $port $protocol)
if [ "$result" != 0 ]; then
local temp=
if [ "$port" -lt 65535 ]; then
temp=$(expr $port + 1)
elif [ "$port" -gt 1 ]; then
temp=$(expr $port - 1)
fi
get_new_port $temp $protocol
else
echo $port
fi
}
check_depends() {
local depends
local tables=${1}
local file_path="/usr/lib/opkg/info"
local file_ext=".control"
[ -d "/lib/apk/packages" ] && file_path="/lib/apk/packages" && file_ext=".list"
if [ "$tables" == "iptables" ]; then
for depends in "iptables-mod-tproxy" "iptables-mod-socket" "iptables-mod-iprange" "iptables-mod-conntrack-extra" "kmod-ipt-nat"; do
[ -s "${file_path}/${depends}${file_ext}" ] || echolog "$tables透明代理基础依赖 $depends 未安装..."
done
else
for depends in "kmod-nft-socket" "kmod-nft-tproxy" "kmod-nft-nat"; do
[ -s "${file_path}/${depends}${file_ext}" ] || echolog "$tables透明代理基础依赖 $depends 未安装..."
done
fi
}
check_ver() {
local version1="$1"
local version2="$2"
local i v1 v1_1 v1_2 v1_3 v2 v2_1 v2_2 v2_3
IFS='.'; set -- $version1; v1_1=${1:-0}; v1_2=${2:-0}; v1_3=${3:-0}
IFS='.'; set -- $version2; v2_1=${1:-0}; v2_2=${2:-0}; v2_3=${3:-0}
IFS=
for i in 1 2 3; do
eval v1=\$v1_$i
eval v2=\$v2_$i
if [ "$v1" -gt "$v2" ]; then
# $1 大于 $2
echo 0
return
elif [ "$v1" -lt "$v2" ]; then
# $1 小于 $2
echo 1
return
fi
done
# $1 等于 $2
echo 255
}
first_type() {
for p in "/bin/$1" "${TMP_BIN_PATH:-/tmp}/$1" "$1"; do
[ -x "$p" ] && echo "$p" && return
done
command -v "$1" 2>/dev/null || command -v "$2" 2>/dev/null
}
eval_set_val() {
for i in $@; do
for j in $i; do
eval $j
done
done
}
eval_unset_val() {
for i in $@; do
for j in $i; do
eval unset j
done
done
}
ln_run() {
local file_func=${1}
local ln_name=${2}
local output=${3}
shift 3;
if [ "${file_func%%/*}" != "${file_func}" ]; then
[ ! -L "${file_func}" ] && {
ln -s "${file_func}" "${TMP_BIN_PATH}/${ln_name}" >/dev/null 2>&1
file_func="${TMP_BIN_PATH}/${ln_name}"
}
[ -x "${file_func}" ] || echolog " - $(readlink ${file_func}) 没有执行权限,无法启动:${file_func} $*"
fi
#echo "${file_func} $*" >&2
[ -n "${file_func}" ] || echolog " - 找不到 ${ln_name},无法启动..."
[ "${output}" != "/dev/null" ] && [ "${ln_name}" != "chinadns-ng" ] && {
local persist_log_path=$(config_t_get global persist_log_path)
local sys_log=$(config_t_get global sys_log "0")
}
if [ -z "$persist_log_path" ] && [ "$sys_log" != "1" ]; then
${file_func:-echolog " - ${ln_name}"} "$@" >${output} 2>&1 &
else
[ "${output: -1, -7}" == "TCP.log" ] && local protocol="TCP"
[ "${output: -1, -7}" == "UDP.log" ] && local protocol="UDP"
if [ -n "${persist_log_path}" ]; then
mkdir -p ${persist_log_path}
local log_file=${persist_log_path}/passwall_${protocol}_${ln_name}_$(date '+%F').log
echolog "记录到持久性日志文件:${log_file}"
${file_func:-echolog " - ${ln_name}"} "$@" >> ${log_file} 2>&1 &
sys_log=0
fi
if [ "${sys_log}" == "1" ]; then
echolog "记录 ${ln_name}_${protocol} 到系统日志"
${file_func:-echolog " - ${ln_name}"} "$@" 2>&1 | logger -t PASSWALL_${protocol}_${ln_name} &
fi
fi
process_count=$(ls $TMP_SCRIPT_FUNC_PATH | wc -l)
process_count=$((process_count + 1))
echo "${file_func:-echolog " - ${ln_name}"} $@ >${output}" > $TMP_SCRIPT_FUNC_PATH/$process_count
}
lua_api() {
local func=${1}
[ -z "${func}" ] && {
echo "nil"
return
}
echo $(lua -e "local api = require 'luci.passwall.api' print(api.${func})")
}
parse_doh() {
local __doh=$1 __url_var=$2 __host_var=$3 __port_var=$4 __bootstrap_var=$5
__doh=$(echo -e "$__doh" | tr -d ' \t\n')
local __url=${__doh%%,*}
local __bootstrap=${__doh#*,}
local __host_port=$(lua_api "get_domain_from_url(\"${__url}\")")
local __host __port
if echo "${__host_port}" | grep -q '^\[.*\]:[0-9]\+$'; then
__host=${__host_port%%]:*}]
__port=${__host_port##*:}
elif echo "${__host_port}" | grep -q ':[0-9]\+$'; then
__host=${__host_port%:*}
__port=${__host_port##*:}
else
__host=${__host_port}
__port=443
fi
__host=${__host#[}
__host=${__host%]}
if [ "$(lua_api "is_ip(\"${__host}\")")" = "true" ]; then
__bootstrap=${__host}
fi
__bootstrap=${__bootstrap#[}
__bootstrap=${__bootstrap%]}
eval "${__url_var}='${__url}' ${__host_var}='${__host}' ${__port_var}='${__port}' ${__bootstrap_var}='${__bootstrap}'"
}
get_geoip() {
local geoip_code="$1"
local geoip_type_flag=""
local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat"
[ -s "$geoip_path" ] || { echo ""; return 1; }
case "$2" in
"ipv4") geoip_type_flag="-ipv6=false" ;;
"ipv6") geoip_type_flag="-ipv4=false" ;;
esac
if type geoview &> /dev/null; then
geoview -input "$geoip_path" -list "$geoip_code" $geoip_type_flag -lowmem=true
return 0
else
echo ""
return 1
fi
}
set_cache_var() {
local key="${1}"
shift 1
local val="$@"
[ -n "${key}" ] && [ -n "${val}" ] && {
sed -i "/${key}=/d" $TMP_PATH/var >/dev/null 2>&1
echo "${key}=\"${val}\"" >> $TMP_PATH/var
eval ${key}=\"${val}\"
}
}
get_cache_var() {
local key="${1}"
[ -n "${key}" ] && [ -s "$TMP_PATH/var" ] && {
echo $(cat $TMP_PATH/var | grep "^${key}=" | awk -F '=' '{print $2}' | tail -n 1 | awk -F'"' '{print $2}')
}
}
eval_cache_var() {
[ -s "$TMP_PATH/var" ] && eval $(cat "$TMP_PATH/var")
}
has_1_65535() {
local val="$1"
val=${val//:/-}
case ",$val," in
*,1-65535,*) return 0 ;;
*) return 1 ;;
esac
}
run_ipt2socks() {
local flag proto tcp_tproxy local_port socks_address socks_port socks_username socks_password log_file
local _extra_param=""
eval_set_val $@
[ -n "$log_file" ] || log_file="/dev/null"
socks_address=$(get_host_ip "ipv4" ${socks_address})
[ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} -a $socks_username -k $socks_password"
[ -n "$tcp_tproxy" ] || _extra_param="${_extra_param} -R"
case "$proto" in
UDP)
flag="${flag}_UDP"
_extra_param="${_extra_param} -U"
;;
TCP)
flag="${flag}_TCP"
_extra_param="${_extra_param} -T"
;;
*)
flag="${flag}_TCP_UDP"
;;
esac
_extra_param="${_extra_param} -n 65535 -v"
ln_run "$(first_type ipt2socks)" "ipt2socks_${flag}" $log_file -l $local_port -b 0.0.0.0 -s $socks_address -p $socks_port ${_extra_param}
}
run_singbox() {
local flag type node tcp_redir_port tcp_proxy_way udp_redir_port socks_address socks_port socks_username socks_password http_address http_port http_username http_password
local dns_listen_port direct_dns_query_strategy direct_dns_port direct_dns_udp_server direct_dns_tcp_server remote_dns_protocol remote_dns_udp_server remote_dns_tcp_server remote_dns_doh remote_dns_client_ip remote_fakedns remote_dns_query_strategy dns_cache dns_socks_address dns_socks_port
local loglevel log_file config_file server_host server_port no_run
local _extra_param=""
eval_set_val $@
[ -z "$type" ] && {
local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z')
if [ "$type" != "sing-box" ]; then
bin=$(first_type $(config_t_get global_app sing_box_file) sing-box)
[ -n "$bin" ] && type="sing-box"
fi
}
[ -z "$type" ] && return 1
[ -n "$log_file" ] || local log_file="/dev/null"
_extra_param="${_extra_param} -log 1 -logfile ${log_file}"
if [ "$log_file" = "/dev/null" ]; then
_extra_param="${_extra_param} -log 0"
else
_extra_param="${_extra_param} -log 1 -logfile ${log_file}"
fi
[ -z "$loglevel" ] && local loglevel=$(config_t_get global loglevel "warn")
[ "$loglevel" = "warning" ] && loglevel="warn"
_extra_param="${_extra_param} -loglevel $loglevel"
[ -n "$flag" ] && _extra_param="${_extra_param} -flag $flag"
[ -n "$node" ] && _extra_param="${_extra_param} -node $node"
[ -n "$server_host" ] && _extra_param="${_extra_param} -server_host $server_host"
[ -n "$server_port" ] && _extra_param="${_extra_param} -server_port $server_port"
[ -n "$tcp_redir_port" ] && _extra_param="${_extra_param} -tcp_redir_port $tcp_redir_port"
[ -n "$tcp_proxy_way" ] && _extra_param="${_extra_param} -tcp_proxy_way $tcp_proxy_way"
[ -n "$udp_redir_port" ] && _extra_param="${_extra_param} -udp_redir_port $udp_redir_port"
[ -n "$socks_address" ] && _extra_param="${_extra_param} -local_socks_address $socks_address"
[ -n "$socks_port" ] && _extra_param="${_extra_param} -local_socks_port $socks_port"
[ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} -local_socks_username $socks_username -local_socks_password $socks_password"
[ -n "$http_address" ] && _extra_param="${_extra_param} -local_http_address $http_address"
[ -n "$http_port" ] && _extra_param="${_extra_param} -local_http_port $http_port"
[ -n "$http_username" ] && [ -n "$http_password" ] && _extra_param="${_extra_param} -local_http_username $http_username -local_http_password $http_password"
[ -n "$dns_socks_address" ] && [ -n "$dns_socks_port" ] && _extra_param="${_extra_param} -dns_socks_address ${dns_socks_address} -dns_socks_port ${dns_socks_port}"
[ -n "$dns_listen_port" ] && _extra_param="${_extra_param} -dns_listen_port ${dns_listen_port}"
[ -n "$dns_cache" ] && _extra_param="${_extra_param} -dns_cache ${dns_cache}"
if [ -n "$direct_dns_udp_server" ]; then
direct_dns_port=$(echo ${direct_dns_udp_server} | awk -F '#' '{print $2}')
_extra_param="${_extra_param} -direct_dns_udp_server $(echo ${direct_dns_udp_server} | awk -F '#' '{print $1}')"
elif [ -n "$direct_dns_tcp_server" ]; then
direct_dns_port=$(echo ${direct_dns_tcp_server} | awk -F '#' '{print $2}')
_extra_param="${_extra_param} -direct_dns_tcp_server $(echo ${direct_dns_tcp_server} | awk -F '#' '{print $1}')"
else
local local_dns=$(echo -n $(echo "${LOCAL_DNS}" | sed "s/,/\n/g" | head -n1) | tr " " ",")
_extra_param="${_extra_param} -direct_dns_udp_server $(echo ${local_dns} | awk -F '#' '{print $1}')"
direct_dns_port=$(echo ${local_dns} | awk -F '#' '{print $2}')
fi
_extra_param="${_extra_param} -direct_dns_port ${direct_dns_port:-53}"
direct_dns_query_strategy=${direct_dns_query_strategy:-UseIP}
_extra_param="${_extra_param} -direct_dns_query_strategy ${direct_dns_query_strategy}"
[ -n "$remote_dns_query_strategy" ] && _extra_param="${_extra_param} -remote_dns_query_strategy ${remote_dns_query_strategy}"
case "$remote_dns_protocol" in
tcp)
local _dns=$(get_first_dns remote_dns_tcp_server 53 | sed 's/#/:/g')
local _dns_address=$(echo ${_dns} | awk -F ':' '{print $1}')
local _dns_port=$(echo ${_dns} | awk -F ':' '{print $2}')
_extra_param="${_extra_param} -remote_dns_server ${_dns_address} -remote_dns_port ${_dns_port} -remote_dns_tcp_server tcp://${_dns}"
;;
doh)
local _doh_url _doh_host _doh_port _doh_bootstrap
parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap
[ -n "$_doh_bootstrap" ] && _extra_param="${_extra_param} -remote_dns_server ${_doh_bootstrap}"
_extra_param="${_extra_param} -remote_dns_port ${_doh_port} -remote_dns_doh_url ${_doh_url} -remote_dns_doh_host ${_doh_host}"
;;
esac
[ -n "$remote_dns_client_ip" ] && _extra_param="${_extra_param} -remote_dns_client_ip ${remote_dns_client_ip}"
[ "$remote_fakedns" = "1" ] && _extra_param="${_extra_param} -remote_dns_fake 1"
[ -n "$no_run" ] && _extra_param="${_extra_param} -no_run 1"
lua $UTIL_SINGBOX gen_config ${_extra_param} > $config_file
[ -n "$no_run" ] || ln_run "$(first_type $(config_t_get global_app sing_box_file) sing-box)" "sing-box" $log_file run -c "$config_file"
}
run_xray() {
local flag type node tcp_redir_port tcp_proxy_way udp_redir_port socks_address socks_port socks_username socks_password http_address http_port http_username http_password
local dns_listen_port direct_dns_query_strategy direct_dns_port direct_dns_udp_server direct_dns_tcp_server remote_dns_udp_server remote_dns_tcp_server remote_dns_doh remote_dns_client_ip remote_fakedns remote_dns_query_strategy dns_cache dns_socks_address dns_socks_port
local loglevel log_file config_file server_host server_port no_run
local _extra_param=""
eval_set_val $@
[ -z "$type" ] && {
local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z')
if [ "$type" != "xray" ]; then
bin=$(first_type $(config_t_get global_app xray_file) xray)
[ -n "$bin" ] && type="xray"
fi
}
[ -z "$type" ] && return 1
[ -n "$log_file" ] || local log_file="/dev/null"
[ -z "$loglevel" ] && local loglevel=$(config_t_get global loglevel "warning")
[ -n "$flag" ] && _extra_param="${_extra_param} -flag $flag"
[ -n "$node" ] && _extra_param="${_extra_param} -node $node"
[ -n "$server_host" ] && _extra_param="${_extra_param} -server_host $server_host"
[ -n "$server_port" ] && _extra_param="${_extra_param} -server_port $server_port"
[ -n "$tcp_redir_port" ] && _extra_param="${_extra_param} -tcp_redir_port $tcp_redir_port"
[ -n "$tcp_proxy_way" ] && _extra_param="${_extra_param} -tcp_proxy_way $tcp_proxy_way"
[ -n "$udp_redir_port" ] && _extra_param="${_extra_param} -udp_redir_port $udp_redir_port"
[ -n "$socks_address" ] && _extra_param="${_extra_param} -local_socks_address $socks_address"
[ -n "$socks_port" ] && _extra_param="${_extra_param} -local_socks_port $socks_port"
[ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} -local_socks_username $socks_username -local_socks_password $socks_password"
[ -n "$http_address" ] && _extra_param="${_extra_param} -local_http_address $http_address"
[ -n "$http_port" ] && _extra_param="${_extra_param} -local_http_port $http_port"
[ -n "$http_username" ] && [ -n "$http_password" ] && _extra_param="${_extra_param} -local_http_username $http_username -local_http_password $http_password"
[ -n "$dns_socks_address" ] && [ -n "$dns_socks_port" ] && _extra_param="${_extra_param} -dns_socks_address ${dns_socks_address} -dns_socks_port ${dns_socks_port}"
[ -n "$dns_listen_port" ] && _extra_param="${_extra_param} -dns_listen_port ${dns_listen_port}"
if [ -n "$direct_dns_udp_server" ]; then
direct_dns_port=$(echo ${direct_dns_udp_server} | awk -F '#' '{print $2}')
_extra_param="${_extra_param} -direct_dns_udp_server $(echo ${direct_dns_udp_server} | awk -F '#' '{print $1}')"
elif [ -n "$direct_dns_tcp_server" ]; then
direct_dns_port=$(echo ${direct_dns_tcp_server} | awk -F '#' '{print $2}')
_extra_param="${_extra_param} -direct_dns_tcp_server $(echo ${direct_dns_tcp_server} | awk -F '#' '{print $1}')"
else
local local_dns=$(echo -n $(echo "${LOCAL_DNS}" | sed "s/,/\n/g" | head -n1) | tr " " ",")
_extra_param="${_extra_param} -direct_dns_udp_server $(echo ${local_dns} | awk -F '#' '{print $1}')"
direct_dns_port=$(echo ${local_dns} | awk -F '#' '{print $2}')
fi
_extra_param="${_extra_param} -direct_dns_port ${direct_dns_port:-53}"
direct_dns_query_strategy=${direct_dns_query_strategy:-UseIP}
_extra_param="${_extra_param} -direct_dns_query_strategy ${direct_dns_query_strategy}"
[ -n "$remote_dns_query_strategy" ] && _extra_param="${_extra_param} -remote_dns_query_strategy ${remote_dns_query_strategy}"
[ -n "$remote_dns_client_ip" ] && _extra_param="${_extra_param} -remote_dns_client_ip ${remote_dns_client_ip}"
[ "$remote_fakedns" = "1" ] && _extra_param="${_extra_param} -remote_dns_fake 1"
[ -n "$dns_cache" ] && _extra_param="${_extra_param} -dns_cache ${dns_cache}"
[ -n "${remote_dns_tcp_server}" ] && {
local _dns=$(get_first_dns remote_dns_tcp_server 53 | sed 's/#/:/g')
local _dns_address=$(echo ${_dns} | awk -F ':' '{print $1}')
local _dns_port=$(echo ${_dns} | awk -F ':' '{print $2}')
_extra_param="${_extra_param} -remote_dns_tcp_server ${_dns_address} -remote_dns_tcp_port ${_dns_port}"
}
[ -n "${remote_dns_doh}" ] && {
local _doh_url _doh_host _doh_port _doh_bootstrap
parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap
[ -n "$_doh_bootstrap" ] && _extra_param="${_extra_param} -remote_dns_doh_ip ${_doh_bootstrap}"
_extra_param="${_extra_param} -remote_dns_doh_port ${_doh_port} -remote_dns_doh_url ${_doh_url} -remote_dns_doh_host ${_doh_host}"
}
_extra_param="${_extra_param} -loglevel $loglevel"
[ -n "$no_run" ] && _extra_param="${_extra_param} -no_run 1"
lua $UTIL_XRAY gen_config ${_extra_param} > $config_file
[ -n "$no_run" ] || ln_run "$(first_type $(config_t_get global_app ${type}_file) ${type})" ${type} $log_file run -c "$config_file"
}
run_dns2socks() {
local flag socks socks_address socks_port socks_username socks_password listen_address listen_port dns cache log_file
local _extra_param=""
eval_set_val $@
[ -n "$flag" ] && flag="_${flag}"
[ -n "$log_file" ] || log_file="/dev/null"
dns=$(get_first_dns dns 53 | sed 's/#/:/g')
[ -n "$socks" ] && {
socks="${socks//#/:}"
socks_address=$(echo $socks | awk -F ':' '{print $1}')
socks_port=$(echo $socks | awk -F ':' '{print $2}')
}
[ -n "$socks_username" ] && [ -n "$socks_password" ] && _extra_param="${_extra_param} /u $socks_username /p $socks_password"
[ -z "$cache" ] && cache=1
[ "$cache" = "0" ] && _extra_param="${_extra_param} /d"
ln_run "$(first_type dns2socks)" "dns2socks${flag}" $log_file ${_extra_param} "${socks_address}:${socks_port}" "${dns}" "${listen_address}:${listen_port}"
}
run_chinadns_ng() {
local _flag _listen_port _dns_local _dns_trust _no_ipv6_trust _use_direct_list _use_proxy_list _gfwlist _chnlist _default_mode _default_tag _no_logic_log _tcp_node _remote_fakedns
local _extra_param=""
eval_set_val $@
local _CONF_FILE=$TMP_ACL_PATH/$_flag/chinadns_ng.conf
local _LOG_FILE="/dev/null"
[ "$(config_t_get global log_chinadns_ng "0")" == "1" ] && _LOG_FILE=$TMP_ACL_PATH/$_flag/chinadns_ng.log
_extra_param="-FLAG ${_flag} -TCP_NODE ${_tcp_node} -LISTEN_PORT ${_listen_port} -DNS_LOCAL ${_dns_local} -DNS_TRUST ${_dns_trust}"
_extra_param="${_extra_param} -USE_DIRECT_LIST ${_use_direct_list} -USE_PROXY_LIST ${_use_proxy_list} -USE_BLOCK_LIST ${_use_block_list}"
_extra_param="${_extra_param} -GFWLIST ${_gfwlist} -CHNLIST ${_chnlist} -NO_IPV6_TRUST ${_no_ipv6_trust} -DEFAULT_MODE ${_default_mode}"
_extra_param="${_extra_param} -DEFAULT_TAG ${_default_tag} -NFTFLAG ${nftflag} -NO_LOGIC_LOG ${_no_logic_log} -REMOTE_FAKEDNS ${_remote_fakedns}"
_extra_param="${_extra_param} -LOG_FILE ${_LOG_FILE}"
lua $APP_PATH/helper_chinadns_add.lua ${_extra_param} > ${_CONF_FILE}
ln_run "$(first_type chinadns-ng)" chinadns-ng "${_LOG_FILE}" -C ${_CONF_FILE}
}
run_socks() {
local flag node bind socks_port config_file http_port http_config_file relay_port log_file no_run
eval_set_val $@
[ -n "$config_file" ] && [ -z "$(echo ${config_file} | grep $TMP_PATH)" ] && config_file=$TMP_PATH/$config_file
[ -n "$http_port" ] || http_port=0
[ -n "$http_config_file" ] && [ -z "$(echo ${http_config_file} | grep $TMP_PATH)" ] && http_config_file=$TMP_PATH/$http_config_file
if [ -n "$log_file" ] && [ -z "$(echo ${log_file} | grep $TMP_PATH)" ]; then
log_file=$TMP_PATH/$log_file
else
log_file="/dev/null"
fi
local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z')
local remarks=$(config_n_get $node remarks)
local server_host=$(config_n_get $node address)
local server_port=$(config_n_get $node port)
[ -n "$relay_port" ] && {
server_host="127.0.0.1"
server_port=$relay_port
}
local error_msg tmp
if [ -n "$server_host" ] && [ -n "$server_port" ]; then
check_host $server_host
[ $? != 0 ] && {
echolog " - Socks节点:[$remarks]${server_host} 是非法的服务器地址,无法启动!"
return 1
}
tmp="${server_host}:${server_port}"
else
error_msg="某种原因,此 Socks 服务的相关配置已失联,启动中止!"
fi
if [ "$type" == "sing-box" ] || [ "$type" == "xray" ]; then
local protocol=$(config_n_get $node protocol)
if [ "$protocol" == "_balancing" ] || [ "$protocol" == "_shunt" ] || [ "$protocol" == "_iface" ] || [ "$protocol" == "_urltest" ]; then
unset error_msg
fi
fi
[ -n "${error_msg}" ] && {
[ "$bind" != "127.0.0.1" ] && echolog " - Socks节点:[$remarks]${tmp},启动中止 ${bind}:${socks_port} ${error_msg}"
return 1
}
[ "$bind" != "127.0.0.1" ] && echolog " - Socks节点:[$remarks]${tmp},启动 ${bind}:${socks_port}"
case "$type" in
socks)
local _socks_address=$(config_n_get $node address)
local _socks_port=$(config_n_get $node port)
local _socks_username=$(config_n_get $node username)
local _socks_password=$(config_n_get $node password)
[ "$http_port" != "0" ] && {
http_flag=1
config_file="${config_file//SOCKS/HTTP_SOCKS}"
local _extra_param="-local_http_address $bind -local_http_port $http_port"
}
local bin=$(first_type $(config_t_get global_app sing_box_file) sing-box)
if [ -n "$bin" ]; then
type="sing-box"
lua $UTIL_SINGBOX gen_proto_config -local_socks_address $bind -local_socks_port $socks_port ${_extra_param} -server_proto socks -server_address ${_socks_address} -server_port ${_socks_port} -server_username ${_socks_username} -server_password ${_socks_password} > $config_file
ln_run "$bin" ${type} $log_file run -c "$config_file"
else
bin=$(first_type $(config_t_get global_app xray_file) xray)
[ -n "$bin" ] && {
type="xray"
lua $UTIL_XRAY gen_proto_config -local_socks_address $bind -local_socks_port $socks_port ${_extra_param} -server_proto socks -server_address ${_socks_address} -server_port ${_socks_port} -server_username ${_socks_username} -server_password ${_socks_password} > $config_file
ln_run "$bin" ${type} $log_file run -c "$config_file"
}
fi
;;
sing-box)
[ "$http_port" != "0" ] && {
http_flag=1
config_file="${config_file//SOCKS/HTTP_SOCKS}"
local _args="http_address=$bind http_port=$http_port"
}
[ -n "$relay_port" ] && _args="${_args} server_host=$server_host server_port=$server_port"
[ -n "$no_run" ] && _args="${_args} no_run=1"
run_singbox flag=$flag node=$node socks_address=$bind socks_port=$socks_port config_file=$config_file log_file=$log_file ${_args}
;;
xray)
[ "$http_port" != "0" ] && {
http_flag=1
config_file="${config_file//SOCKS/HTTP_SOCKS}"
local _args="http_address=$bind http_port=$http_port"
}
[ -n "$relay_port" ] && _args="${_args} server_host=$server_host server_port=$server_port"
[ -n "$no_run" ] && _args="${_args} no_run=1"
run_xray flag=$flag node=$node socks_address=$bind socks_port=$socks_port config_file=$config_file log_file=$log_file ${_args}
;;
trojan*)
lua $UTIL_TROJAN gen_config -node $node -run_type client -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file
[ -n "$no_run" ] || ln_run "$(first_type ${type})" "${type}" $log_file -c "$config_file"
;;
naiveproxy)
lua $UTIL_NAIVE gen_config -node $node -run_type socks -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file
[ -n "$no_run" ] || ln_run "$(first_type naive)" naive $log_file "$config_file"
;;
ssr)
lua $UTIL_SS gen_config -node $node -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file
[ -n "$no_run" ] || ln_run "$(first_type ssr-local)" "ssr-local" $log_file -c "$config_file" -v -u
;;
ss)
[ -n "$no_run" ] || {
local plugin_sh="${config_file%.json}_plugin.sh"
local _extra_param="-plugin_sh $plugin_sh"
}
lua $UTIL_SS gen_config -node $node -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port -mode tcp_and_udp ${_extra_param} > $config_file
[ -n "$no_run" ] || ln_run "$(first_type ss-local)" "ss-local" $log_file -c "$config_file" -v
;;
ss-rust)
local _extra_param
[ "$http_port" != "0" ] && {
http_flag=1
config_file="${config_file//SOCKS/HTTP_SOCKS}"
_extra_param="-local_http_address $bind -local_http_port $http_port"
}
[ -n "$no_run" ] || {
local plugin_sh="${config_file%.json}_plugin.sh"
_extra_param="${_extra_param:+$_extra_param }-plugin_sh $plugin_sh"
}
lua $UTIL_SS gen_config -node $node -local_socks_address $bind -local_socks_port $socks_port -server_host $server_host -server_port $server_port ${_extra_param} > $config_file
[ -n "$no_run" ] || ln_run "$(first_type sslocal)" "sslocal" $log_file -c "$config_file" -v
;;
hysteria2)
[ "$http_port" != "0" ] && {
http_flag=1
config_file="${config_file//SOCKS/HTTP_SOCKS}"
local _extra_param="-local_http_address $bind -local_http_port $http_port"
}
lua $UTIL_HYSTERIA2 gen_config -node $node -local_socks_address $bind -local_socks_port $socks_port -server_host $server_host -server_port $server_port ${_extra_param} > $config_file
[ -n "$no_run" ] || ln_run "$(first_type $(config_t_get global_app hysteria_file))" "hysteria" $log_file -c "$config_file" client
;;
tuic)
lua $UTIL_TUIC gen_config -node $node -local_addr $bind -local_port $socks_port -server_host $server_host -server_port $server_port > $config_file
[ -n "$no_run" ] || ln_run "$(first_type tuic-client)" "tuic-client" $log_file -c "$config_file"
;;
esac
set_cache_var "node_${node}_socks_port" "${socks_port}"
# http to socks
[ -z "$http_flag" ] && [ "$http_port" != "0" ] && [ -n "$http_config_file" ] && [ "$type" != "sing-box" ] && [ "$type" != "xray" ] && [ "$type" != "socks" ] && {
local bin=$(first_type $(config_t_get global_app sing_box_file) sing-box)
if [ -n "$bin" ]; then
type="sing-box"
lua $UTIL_SINGBOX gen_proto_config -local_http_address $bind -local_http_port $http_port -server_proto socks -server_address "127.0.0.1" -server_port $socks_port -server_username $_username -server_password $_password > $http_config_file
[ -n "$no_run" ] || ln_run "$bin" ${type} /dev/null run -c "$http_config_file"
else
bin=$(first_type $(config_t_get global_app xray_file) xray)
[ -n "$bin" ] && type="xray"
[ -z "$type" ] && return 1
lua $UTIL_XRAY gen_proto_config local_http_address $bind -local_http_port $http_port -server_proto socks -server_address "127.0.0.1" -server_port $socks_port -server_username $_username -server_password $_password > $http_config_file
[ -n "$no_run" ] || ln_run "$bin" ${type} /dev/null run -c "$http_config_file"
fi
}
unset http_flag
[ -z "$no_run" ] && [ "${server_host}" != "127.0.0.1" ] && [ "$type" != "sing-box" ] && [ "$type" != "xray" ] && echo "${node}" >> $TMP_PATH/direct_node_list
}
run_redir() {
local node proto bind local_port config_file log_file
eval_set_val $@
local tcp_node_socks_flag tcp_node_http_flag
[ -n "$config_file" ] && [ -z "$(echo ${config_file} | grep $TMP_PATH)" ] && config_file=${GLOBAL_ACL_PATH}/${config_file}
if [ -n "$log_file" ] && [ -z "$(echo ${log_file} | grep $TMP_PATH)" ]; then
log_file=${GLOBAL_ACL_PATH}/${log_file}
else
log_file="/dev/null"
fi
local proto=$(echo $proto | tr 'A-Z' 'a-z')
local PROTO=$(echo $proto | tr 'a-z' 'A-Z')
local type=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z')
local enable_log=$(config_t_get global log_${proto} 1)
[ "$enable_log" != "1" ] && log_file="/dev/null"
local remarks=$(config_n_get $node remarks)
local server_host=$(config_n_get $node address)
local port=$(config_n_get $node port)
[ -n "$server_host" ] && [ -n "$port" ] && {
check_host $server_host
[ $? != 0 ] && {
echolog "${PROTO}节点:[$remarks]${server_host} 是非法的服务器地址,无法启动!"
return 1
}
}
[ "$bind" != "127.0.0.1" ] && echolog "${PROTO}节点:[$remarks],监听端口:$local_port"
case "$PROTO" in
UDP)
case "$type" in
socks)
local _socks_address=$(config_n_get $node address)
local _socks_port=$(config_n_get $node port)
local _socks_username=$(config_n_get $node username)
local _socks_password=$(config_n_get $node password)
run_ipt2socks flag=default proto=UDP local_port=${local_port} socks_address=${_socks_address} socks_port=${_socks_port} socks_username=${_socks_username} socks_password=${_socks_password} log_file=${log_file}
;;
sing-box)
local protocol=$(config_n_get $node protocol)
[ "$protocol" = "_shunt" ] && {
local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat"
local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat"
if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then
echolog "* 缺少Geo规则文件,UDP Sing-Box分流节点无法正常使用!"
fi
}
run_singbox flag=UDP node=$node udp_redir_port=$local_port config_file=$config_file log_file=$log_file
;;
xray)
local protocol=$(config_n_get $node protocol)
[ "$protocol" = "_shunt" ] && {
local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat"
local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat"
if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then
echolog "* 缺少Geo规则文件,UDP Xray分流节点无法正常使用!"
fi
}
run_xray flag=UDP node=$node udp_redir_port=$local_port config_file=$config_file log_file=$log_file
;;
trojan*)
local loglevel=$(config_t_get global trojan_loglevel "2")
lua $UTIL_TROJAN gen_config -node $node -run_type nat -local_addr "0.0.0.0" -local_port $local_port -loglevel $loglevel > $config_file
ln_run "$(first_type ${type})" "${type}" $log_file -c "$config_file"
;;
naiveproxy)
echolog "Naiveproxy不支持UDP转发!"
;;
ssr)
lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port > $config_file
ln_run "$(first_type ssr-redir)" "ssr-redir" $log_file -c "$config_file" -v -U
;;
ss)
local plugin_sh="${config_file%.json}_plugin.sh"
local _extra_param="-plugin_sh $plugin_sh"
lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port -mode udp_only $_extra_param > $config_file
ln_run "$(first_type ss-redir)" "ss-redir" $log_file -c "$config_file" -v
;;
ss-rust)
local plugin_sh="${config_file%.json}_plugin.sh"
local _extra_param="-plugin_sh $plugin_sh"
lua $UTIL_SS gen_config -node $node -local_udp_redir_port $local_port $_extra_param > $config_file
ln_run "$(first_type sslocal)" "sslocal" $log_file -c "$config_file" -v
;;
hysteria2)
lua $UTIL_HYSTERIA2 gen_config -node $node -local_udp_redir_port $local_port > $config_file
ln_run "$(first_type $(config_t_get global_app hysteria_file))" "hysteria" $log_file -c "$config_file" client
;;
tuic)
echolog "TUIC不支持UDP转发!"
;;
esac
;;
TCP)
[ "$TCP_UDP" = "1" ] && echolog "UDP节点:与TCP节点相同"
tcp_node_socks=1
tcp_node_socks_bind_local=$(config_t_get global tcp_node_socks_bind_local 1)
tcp_node_socks_bind="127.0.0.1"
[ "${tcp_node_socks_bind_local}" != "1" ] && tcp_node_socks_bind="0.0.0.0"
tcp_node_socks_port=$(get_new_port $(config_t_get global tcp_node_socks_port 1070))
tcp_node_http_port=$(config_t_get global tcp_node_http_port 0)
[ "$tcp_node_http_port" != "0" ] && tcp_node_http=1
if [ $PROXY_IPV6 == "1" ]; then
echolog "开启实验性IPv6透明代理(TProxy),请确认您的节点及类型支持IPv6!"
fi
if [ "${TCP_PROXY_WAY}" = "redirect" ]; then
can_ipt=$(echo "$REDIRECT_LIST" | grep "$type")
elif [ "${TCP_PROXY_WAY}" = "tproxy" ]; then
can_ipt=$(echo "$TPROXY_LIST" | grep "$type")
fi
[ -z "$can_ipt" ] && type="socks"
case "$type" in
socks)
_socks_flag=1
_socks_address=$(config_n_get $node address)
_socks_port=$(config_n_get $node port)
_socks_username=$(config_n_get $node username)
_socks_password=$(config_n_get $node password)
[ -z "$can_ipt" ] && {
local _config_file=$config_file
_config_file="TCP_SOCKS_${node}.json"
local _port=$(get_new_port 2080)
run_socks flag="TCP" node=$node bind=127.0.0.1 socks_port=${_port} config_file=${_config_file}
_socks_address=127.0.0.1
_socks_port=${_port}
unset _socks_username
unset _socks_password
}
;;
sing-box)
local _flag="TCP"
local _args=""
[ "$tcp_node_socks" = "1" ] && {
tcp_node_socks_flag=1
_args="${_args} socks_address=${tcp_node_socks_bind} socks_port=${tcp_node_socks_port}"
config_file="${config_file//TCP/TCP_SOCKS}"
}
[ "$tcp_node_http" = "1" ] && {
tcp_node_http_flag=1
_args="${_args} http_port=${tcp_node_http_port}"
config_file="${config_file//TCP/TCP_HTTP}"
}
[ "$TCP_UDP" = "1" ] && {
UDP_REDIR_PORT=$local_port
unset UDP_NODE
_flag="TCP_UDP"
_args="${_args} udp_redir_port=${UDP_REDIR_PORT}"
config_file="${config_file//TCP/TCP_UDP}"
}
local protocol=$(config_n_get $node protocol)
local default_node=$(config_n_get $node default_node)
local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp)
[ "${DNS_MODE}" != "sing-box" ] && [ "${DNS_MODE}" != "udp" ] && [ "$protocol" = "_shunt" ] && [ "$default_node" = "_direct" ] && {
DNS_MODE="sing-box"
v2ray_dns_mode="tcp"
echolog "* 当前TCP节点采用Sing-Box分流且默认节点为直连,远程DNS过滤模式将默认使用Sing-Box(TCP),防止环回!"
}
[ "$protocol" = "_shunt" ] && {
local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat"
local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat"
if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then
echolog "* 缺少Geo规则文件,TCP Sing-Box分流节点无法正常使用!"
fi
}
[ "${DNS_MODE}" = "sing-box" ] && {
NO_PLUGIN_DNS=1
config_file="${config_file//.json/_DNS.json}"
_args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}"
DNSMASQ_FILTER_PROXY_IPV6=0
local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip)
[ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}"
[ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0"
resolve_dns_port=${NEXT_DNS_LISTEN_PORT}
_args="${_args} dns_listen_port=${resolve_dns_port}"
case "$(config_t_get global direct_dns_mode "auto")" in
udp)
_args="${_args} direct_dns_udp_server=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g')"
;;
tcp)
_args="${_args} direct_dns_tcp_server=$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g')"
;;
esac
_args="${_args} remote_dns_protocol=${v2ray_dns_mode}"
case "$v2ray_dns_mode" in
tcp)
_args="${_args} remote_dns_tcp_server=${REMOTE_DNS}"
resolve_dns_log="Sing-Box DNS(127.0.0.1#${resolve_dns_port}) -> tcp://${REMOTE_DNS}"
;;
doh)
remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query")
_args="${_args} remote_dns_doh=${remote_dns_doh}"
resolve_dns_log="Sing-Box DNS(127.0.0.1#${resolve_dns_port}) -> ${remote_dns_doh}"
;;
esac
local remote_fakedns=$(config_t_get global remote_fakedns 0)
[ "${remote_fakedns}" = "1" ] && {
fakedns=1
_args="${_args} remote_fakedns=1"
resolve_dns_log="${resolve_dns_log} + FakeDNS"
}
NEXT_DNS_LISTEN_PORT=$(expr $NEXT_DNS_LISTEN_PORT + 1)
}
run_singbox flag=$_flag node=$node tcp_redir_port=$local_port tcp_proxy_way=$TCP_PROXY_WAY config_file=$config_file log_file=$log_file ${_args}
;;
xray)
local _flag="TCP"
local _args=""
[ "$tcp_node_socks" = "1" ] && {
tcp_node_socks_flag=1
_args="${_args} socks_address=${tcp_node_socks_bind} socks_port=${tcp_node_socks_port}"
config_file="${config_file//TCP/TCP_SOCKS}"
}
[ "$tcp_node_http" = "1" ] && {
tcp_node_http_flag=1
_args="${_args} http_port=${tcp_node_http_port}"
config_file="${config_file//TCP/TCP_HTTP}"
}
[ "$TCP_UDP" = "1" ] && {
UDP_REDIR_PORT=$local_port
unset UDP_NODE
_flag="TCP_UDP"
_args="${_args} udp_redir_port=${UDP_REDIR_PORT}"
config_file="${config_file//TCP/TCP_UDP}"
}
local protocol=$(config_n_get $node protocol)
local default_node=$(config_n_get $node default_node)
local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp)
[ "${DNS_MODE}" != "xray" ] && [ "${DNS_MODE}" != "udp" ] && [ "$protocol" = "_shunt" ] && [ "$default_node" = "_direct" ] && {
DNS_MODE="xray"
v2ray_dns_mode="tcp"
echolog "* 当前TCP节点采用Xray分流且默认节点为直连,远程DNS过滤模式将默认使用Xray(TCP),防止环回!"
}
[ "$protocol" = "_shunt" ] && {
local geoip_path="${V2RAY_LOCATION_ASSET%*/}/geoip.dat"
local geosite_path="${V2RAY_LOCATION_ASSET%*/}/geosite.dat"
if [ ! -s "$geoip_path" ] || [ ! -s "$geosite_path" ]; then
echolog "* 缺少Geo规则文件,TCP Xray分流节点无法正常使用!"
fi
}
[ "${DNS_MODE}" = "xray" ] && {
NO_PLUGIN_DNS=1
config_file="${config_file//.json/_DNS.json}"
_args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}"
DNSMASQ_FILTER_PROXY_IPV6=0
local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip)
[ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}"
[ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0"
resolve_dns_port=${NEXT_DNS_LISTEN_PORT}
_args="${_args} dns_listen_port=${resolve_dns_port}"
case "$(config_t_get global direct_dns_mode "auto")" in
udp)
_args="${_args} direct_dns_udp_server=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g')"
;;
tcp)
_args="${_args} direct_dns_tcp_server=$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g')"
;;
esac
_args="${_args} remote_dns_tcp_server=${REMOTE_DNS}"
if [ "$v2ray_dns_mode" = "tcp+doh" ]; then
remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query")
_args="${_args} remote_dns_doh=${remote_dns_doh}"
resolve_dns_log="Xray DNS(127.0.0.1#${resolve_dns_port}) -> (${remote_dns_doh})(A/AAAA) + tcp://${REMOTE_DNS}"
else
resolve_dns_log="Xray DNS(127.0.0.1#${resolve_dns_port}) -> tcp://${REMOTE_DNS}"
fi
local remote_fakedns=$(config_t_get global remote_fakedns 0)
[ "${remote_fakedns}" = "1" ] && {
fakedns=1
_args="${_args} remote_fakedns=1"
resolve_dns_log="${resolve_dns_log} + FakeDNS"
}
NEXT_DNS_LISTEN_PORT=$(expr $NEXT_DNS_LISTEN_PORT + 1)
}
run_xray flag=$_flag node=$node tcp_redir_port=$local_port tcp_proxy_way=$TCP_PROXY_WAY config_file=$config_file log_file=$log_file ${_args}
;;
trojan*)
[ "${TCP_PROXY_WAY}" = "tproxy" ] && lua_tproxy_arg="-use_tproxy true"
[ "$TCP_UDP" = "1" ] && {
config_file="${config_file//TCP/TCP_UDP}"
UDP_REDIR_PORT=$TCP_REDIR_PORT
unset UDP_NODE
}
local loglevel=$(config_t_get global trojan_loglevel "2")
lua $UTIL_TROJAN gen_config -node $node -run_type nat -local_addr "0.0.0.0" -local_port $local_port -loglevel $loglevel $lua_tproxy_arg > $config_file
ln_run "$(first_type ${type})" "${type}" $log_file -c "$config_file"
;;
naiveproxy)
lua $UTIL_NAIVE gen_config -node $node -run_type redir -local_addr "0.0.0.0" -local_port $local_port > $config_file
ln_run "$(first_type naive)" naive $log_file "$config_file"
;;
ssr)
[ "${TCP_PROXY_WAY}" = "tproxy" ] && lua_tproxy_arg="-tcp_tproxy true"
[ "$TCP_UDP" = "1" ] && {
config_file="${config_file//TCP/TCP_UDP}"
UDP_REDIR_PORT=$TCP_REDIR_PORT
unset UDP_NODE
_extra_param="-u"
}
lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port $lua_tproxy_arg > $config_file
ln_run "$(first_type ssr-redir)" "ssr-redir" $log_file -c "$config_file" -v ${_extra_param}
;;
ss)
[ "${TCP_PROXY_WAY}" = "tproxy" ] && lua_tproxy_arg="-tcp_tproxy true"
local lua_mode_arg="-mode tcp_only"
[ "$TCP_UDP" = "1" ] && {
config_file="${config_file//TCP/TCP_UDP}"
UDP_REDIR_PORT=$TCP_REDIR_PORT
unset UDP_NODE
lua_mode_arg="-mode tcp_and_udp"
}
local plugin_sh="${config_file%.json}_plugin.sh"
lua_mode_arg="${lua_mode_arg} -plugin_sh $plugin_sh"
lua $UTIL_SS gen_config -node $node -local_addr "0.0.0.0" -local_port $local_port $lua_mode_arg $lua_tproxy_arg > $config_file
ln_run "$(first_type ss-redir)" "ss-redir" $log_file -c "$config_file" -v
;;
ss-rust)
local _extra_param="-local_tcp_redir_port $local_port"
[ "${TCP_PROXY_WAY}" = "tproxy" ] && _extra_param="${_extra_param} -tcp_tproxy true"
[ "$tcp_node_socks" = "1" ] && {
tcp_node_socks_flag=1
config_file="${config_file//TCP/TCP_SOCKS}"
_extra_param="${_extra_param} -local_socks_address ${tcp_node_socks_bind} -local_socks_port ${tcp_node_socks_port}"
}
[ "$tcp_node_http" = "1" ] && {
tcp_node_http_flag=1
config_file="${config_file//TCP/TCP_HTTP}"
_extra_param="${_extra_param} -local_http_port ${tcp_node_http_port}"
}
[ "$TCP_UDP" = "1" ] && {
config_file="${config_file//TCP/TCP_UDP}"
UDP_REDIR_PORT=$TCP_REDIR_PORT
unset UDP_NODE
_extra_param="${_extra_param} -local_udp_redir_port $local_port"
}
local plugin_sh="${config_file%.json}_plugin.sh"
_extra_param="${_extra_param} -plugin_sh $plugin_sh"
lua $UTIL_SS gen_config -node $node ${_extra_param} > $config_file
ln_run "$(first_type sslocal)" "sslocal" $log_file -c "$config_file" -v
;;
hysteria2)
local _extra_param="-local_tcp_redir_port $local_port"
[ "$tcp_node_socks" = "1" ] && {
tcp_node_socks_flag=1
config_file="${config_file//TCP/TCP_SOCKS}"
_extra_param="${_extra_param} -local_socks_address ${tcp_node_socks_bind} -local_socks_port ${tcp_node_socks_port}"
}
[ "$tcp_node_http" = "1" ] && {
tcp_node_http_flag=1
config_file="${config_file//TCP/TCP_HTTP}"
_extra_param="${_extra_param} -local_http_port ${tcp_node_http_port}"
}
[ "$TCP_UDP" = "1" ] && {
config_file="${config_file//TCP/TCP_UDP}"
UDP_REDIR_PORT=$TCP_REDIR_PORT
unset UDP_NODE
_extra_param="${_extra_param} -local_udp_redir_port $local_port"
}
_extra_param="${_extra_param} -tcp_proxy_way ${TCP_PROXY_WAY}"
lua $UTIL_HYSTERIA2 gen_config -node $node ${_extra_param} > $config_file
ln_run "$(first_type $(config_t_get global_app hysteria_file))" "hysteria" $log_file -c "$config_file" client
;;
esac
if [ -n "${_socks_flag}" ]; then
local _flag="TCP"
[ "$TCP_UDP" = "1" ] && {
_flag="TCP_UDP"
UDP_REDIR_PORT=$TCP_REDIR_PORT
unset UDP_NODE
}
local _socks_tproxy=""
[ "${TCP_PROXY_WAY}" = "tproxy" ] && _socks_tproxy="1"
run_ipt2socks flag=default proto=${_flag} tcp_tproxy=${_socks_tproxy} local_port=${local_port} socks_address=${_socks_address} socks_port=${_socks_port} socks_username=${_socks_username} socks_password=${_socks_password} log_file=${log_file}
fi
[ -z "$tcp_node_socks_flag" ] && {
[ "$tcp_node_socks" = "1" ] && {
local config_file="SOCKS_TCP.json"
local log_file="SOCKS_TCP.log"
local http_port=0
local http_config_file="HTTP2SOCKS_TCP.json"
[ "$tcp_node_http" = "1" ] && [ -z "$tcp_node_http_flag" ] && {
http_port=$tcp_node_http_port
}
run_socks flag=TCP node=$node bind=$tcp_node_socks_bind socks_port=$tcp_node_socks_port config_file=$config_file http_port=$http_port http_config_file=$http_config_file
}
}
[ "$tcp_node_socks" = "1" ] && {
TCP_SOCKS_server="127.0.0.1:$tcp_node_socks_port"
set_cache_var "GLOBAL_TCP_SOCKS_server" "${TCP_SOCKS_server}"
}
;;
esac
unset tcp_node_socks_flag tcp_node_http_flag
[ "$type" != "sing-box" ] && [ "$type" != "xray" ] && echo "${node}" >> $TMP_PATH/direct_node_list
[ -n "${redir_port}" ] && set_cache_var "node_${node}_${PROTO}_redir_port" "${local_port}"
}
start_redir() {
local proto=${1}
eval node=\$${proto}_NODE
if [ -n "$node" ]; then
TYPE=$(echo $(config_n_get $node type) | tr 'A-Z' 'a-z')
local config_file="${proto}.json"
local log_file="${proto}.log"
eval current_port=\$${proto}_REDIR_PORT
local port=$(echo $(get_new_port $current_port $proto))
eval ${proto}_REDIR_PORT=$port
run_redir node=$node proto=${proto} bind=0.0.0.0 local_port=$port config_file=$config_file log_file=$log_file
set_cache_var "ACL_GLOBAL_${proto}_node" "${node}"
set_cache_var "ACL_GLOBAL_${proto}_redir_port" "${port}"
[ "$TCP_UDP" = "1" ] && {
set_cache_var "ACL_GLOBAL_UDP_node" "${node}"
set_cache_var "ACL_GLOBAL_UDP_redir_port" "${port}"
}
else
[ "${proto}" = "UDP" ] && [ "$TCP_UDP" = "1" ] && return
echolog "${proto}节点没有选择或为空,不代理${proto}。"
fi
}
start_socks() {
[ "$SOCKS_ENABLED" = "1" ] && {
local ids=$(uci show $CONFIG | grep "=socks" | awk -F '.' '{print $2}' | awk -F '=' '{print $1}')
[ -n "$ids" ] && {
echolog "分析 Socks 服务的节点配置..."
for id in $ids; do
local enabled=$(config_n_get $id enabled 0)
[ "$enabled" == "0" ] && continue
local node=$(config_n_get $id node)
[ -z "$node" ] && continue
local bind_local=$(config_n_get $id bind_local 0)
local bind="0.0.0.0"
[ "$bind_local" = "1" ] && bind="127.0.0.1"
local port=$(config_n_get $id port)
local config_file="SOCKS_${id}.json"
local log_file="SOCKS_${id}.log"
local log=$(config_n_get $id log 1)
[ "$log" == "0" ] && log_file=""
local http_port=$(config_n_get $id http_port 0)
local http_config_file="HTTP2SOCKS_${id}.json"
run_socks flag=$id node=$node bind=$bind socks_port=$port config_file=$config_file http_port=$http_port http_config_file=$http_config_file log_file=$log_file
set_cache_var "socks_${id}" "$node"
#自动切换逻辑
local enable_autoswitch=$(config_n_get $id enable_autoswitch 0)
[ "$enable_autoswitch" = "1" ] && $APP_PATH/socks_auto_switch.sh ${id} > /dev/null 2>&1 &
done
}
}
}
socks_node_switch() {
local flag new_node
eval_set_val $@
[ -n "$flag" ] && [ -n "$new_node" ] && {
local prefix pf filename
# 结束 SS 插件进程
for prefix in "" "HTTP_"; do
pf="$TMP_PATH/${prefix}SOCKS_${flag}_plugin.pid"
[ -s "$pf" ] && kill -9 "$(head -n1 "$pf")" >/dev/null 2>&1
done
pgrep -af "$TMP_BIN_PATH" | awk -v P1="${flag}" 'BEGIN{IGNORECASE=1}$0~P1 && !/acl\/|acl_/{print $1}' | xargs kill -9 >/dev/null 2>&1
for prefix in "" "HTTP_" "HTTP2"; do
rm -rf "$TMP_PATH/${prefix}SOCKS_${flag}"*
done
for filename in $(ls ${TMP_SCRIPT_FUNC_PATH}); do
cmd=$(cat ${TMP_SCRIPT_FUNC_PATH}/${filename})
[ -n "$(echo $cmd | grep "${flag}")" ] && rm -f ${TMP_SCRIPT_FUNC_PATH}/${filename}
done
local bind_local=$(config_n_get $flag bind_local 0)
local bind="0.0.0.0"
[ "$bind_local" = "1" ] && bind="127.0.0.1"
local port=$(config_n_get $flag port)
local config_file="SOCKS_${flag}.json"
local log_file="SOCKS_${flag}.log"
local log=$(config_n_get $flag log 1)
[ "$log" == "0" ] && log_file=""
local http_port=$(config_n_get $flag http_port 0)
local http_config_file="HTTP2SOCKS_${flag}.json"
LOG_FILE="/dev/null"
run_socks flag=$flag node=$new_node bind=$bind socks_port=$port config_file=$config_file http_port=$http_port http_config_file=$http_config_file log_file=$log_file
set_cache_var "socks_${flag}" "$new_node"
local USE_TABLES=$(get_cache_var "USE_TABLES")
[ -n "$USE_TABLES" ] && source $APP_PATH/${USE_TABLES}.sh filter_direct_node_list
}
}
clean_log() {
logsnum=$(cat $LOG_FILE 2>/dev/null | wc -l)
[ "$logsnum" -gt 1000 ] && {
echo "" > $LOG_FILE
echolog "日志文件过长,清空处理!"
}
}
clean_crontab() {
[ -f "/tmp/lock/${CONFIG}_cron.lock" ] && return
touch /etc/crontabs/root
#sed -i "/${CONFIG}/d" /etc/crontabs/root >/dev/null 2>&1
sed -i "/$(echo "/etc/init.d/${CONFIG}" | sed 's#\/#\\\/#g')/d" /etc/crontabs/root >/dev/null 2>&1
sed -i "/$(echo "lua ${APP_PATH}/rule_update.lua log" | sed 's#\/#\\\/#g')/d" /etc/crontabs/root >/dev/null 2>&1
sed -i "/$(echo "lua ${APP_PATH}/subscribe.lua start" | sed 's#\/#\\\/#g')/d" /etc/crontabs/root >/dev/null 2>&1
pgrep -af "${CONFIG}/" | awk '/tasks\.sh/{print $1}' | xargs kill -9 >/dev/null 2>&1
rm -rf /tmp/lock/${CONFIG}_tasks.lock
}
start_crontab() {
if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then
start_daemon=$(config_t_get global_delay start_daemon 0)
[ "$start_daemon" = "1" ] && $APP_PATH/monitor.sh > /dev/null 2>&1 &
fi
[ -f "/tmp/lock/${CONFIG}_cron.lock" ] && {
rm -rf "/tmp/lock/${CONFIG}_cron.lock"
echolog "当前为计划任务自动运行,不重新配置定时任务。"
return
}
clean_crontab
[ "$ENABLED" != 1 ] && {
/etc/init.d/cron restart
return
}
stop_week_mode=$(config_t_get global_delay stop_week_mode)
stop_time_mode=$(config_t_get global_delay stop_time_mode)
if [ -n "$stop_week_mode" ]; then
local t="0 $stop_time_mode * * $stop_week_mode"
[ "$stop_week_mode" = "7" ] && t="0 $stop_time_mode * * *"
if [ "$stop_week_mode" = "8" ]; then
update_loop=1
else
echo "$t /etc/init.d/$CONFIG stop > /dev/null 2>&1 &" >>/etc/crontabs/root
fi
echolog "配置定时任务:自动关闭服务。"
fi
start_week_mode=$(config_t_get global_delay start_week_mode)
start_time_mode=$(config_t_get global_delay start_time_mode)
if [ -n "$start_week_mode" ]; then
local t="0 $start_time_mode * * $start_week_mode"
[ "$start_week_mode" = "7" ] && t="0 $start_time_mode * * *"
if [ "$start_week_mode" = "8" ]; then
update_loop=1
else
echo "$t /etc/init.d/$CONFIG start > /dev/null 2>&1 &" >>/etc/crontabs/root
fi
echolog "配置定时任务:自动开启服务。"
fi
restart_week_mode=$(config_t_get global_delay restart_week_mode)
restart_time_mode=$(config_t_get global_delay restart_time_mode)
if [ -n "$restart_week_mode" ]; then
local t="0 $restart_time_mode * * $restart_week_mode"
[ "$restart_week_mode" = "7" ] && t="0 $restart_time_mode * * *"
if [ "$restart_week_mode" = "8" ]; then
update_loop=1
else
echo "$t /etc/init.d/$CONFIG restart > /dev/null 2>&1 &" >>/etc/crontabs/root
fi
echolog "配置定时任务:自动重启服务。"
fi
autoupdate=$(config_t_get global_rules auto_update)
weekupdate=$(config_t_get global_rules week_update)
dayupdate=$(config_t_get global_rules time_update)
if [ "$autoupdate" = "1" ]; then
local t="0 $dayupdate * * $weekupdate"
[ "$weekupdate" = "7" ] && t="0 $dayupdate * * *"
if [ "$weekupdate" = "8" ]; then
update_loop=1
else
echo "$t lua $APP_PATH/rule_update.lua log all cron > /dev/null 2>&1 &" >>/etc/crontabs/root
fi
echolog "配置定时任务:自动更新规则。"
fi
TMP_SUB_PATH=$TMP_PATH/sub_crontabs
mkdir -p $TMP_SUB_PATH
for item in $(uci show ${CONFIG} | grep "=subscribe_list" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do
if [ "$(config_n_get $item auto_update 0)" = "1" ]; then
cfgid=$(uci show ${CONFIG}.$item | head -n 1 | cut -d '.' -sf 2 | cut -d '=' -sf 1)
remark=$(config_n_get $item remark)
week_update=$(config_n_get $item week_update)
time_update=$(config_n_get $item time_update)
echo "$cfgid" >> $TMP_SUB_PATH/${week_update}_${time_update}
echolog "配置定时任务:自动更新【$remark】订阅。"
fi
done
[ -d "${TMP_SUB_PATH}" ] && {
for name in $(ls ${TMP_SUB_PATH}); do
week_update=$(echo $name | awk -F '_' '{print $1}')
time_update=$(echo $name | awk -F '_' '{print $2}')
cfgids=$(echo -n $(cat ${TMP_SUB_PATH}/${name}) | sed 's# #,#g')
local t="0 $time_update * * $week_update"
[ "$week_update" = "7" ] && t="0 $time_update * * *"
if [ "$week_update" = "8" ]; then
update_loop=1
else
echo "$t lua $APP_PATH/subscribe.lua start $cfgids cron > /dev/null 2>&1 &" >>/etc/crontabs/root
fi
done
rm -rf $TMP_SUB_PATH
}
if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then
[ "$update_loop" = "1" ] && {
$APP_PATH/tasks.sh > /dev/null 2>&1 &
echolog "自动更新:启动循环更新进程。"
}
else
echolog "运行于非代理模式,仅允许服务启停的定时任务。"
fi
/etc/init.d/cron restart
}
stop_crontab() {
[ -f "/tmp/lock/${CONFIG}_cron.lock" ] && return
clean_crontab
/etc/init.d/cron restart
#echolog "清除定时执行命令。"
}
start_dns() {
echolog "DNS域名解析:"
local china_ng_local_dns=$(IFS=','; set -- $LOCAL_DNS; [ "${1%%[#:]*}" = "127.0.0.1" ] && echo "$1" || ([ -n "$2" ] && echo "$1,$2" || echo "$1"))
local sing_box_local_dns=
local direct_dns_mode=$(config_t_get global direct_dns_mode "auto")
#获取访问控制节点所使用的DNS分流模式
local ACL_RULE_DNSMASQ=0
for acl_section in $(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1); do
if [ "$(config_n_get $acl_section enabled)" = "1" ] && \
[ "$(config_n_get $acl_section dns_shunt)" = "dnsmasq" ]; then
ACL_RULE_DNSMASQ=1
break
fi
done
case "$direct_dns_mode" in
udp)
LOCAL_DNS=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g')
china_ng_local_dns=${LOCAL_DNS}
sing_box_local_dns="direct_dns_udp_server=${LOCAL_DNS}"
;;
tcp)
local DIRECT_DNS=$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g')
china_ng_local_dns="tcp://${DIRECT_DNS}"
sing_box_local_dns="direct_dns_tcp_server=${DIRECT_DNS}"
#当全局(包括访问控制节点)开启chinadns-ng时,不启动新进程。
[ "$DNS_SHUNT" != "chinadns-ng" ] || [ "$ACL_RULE_DNSMASQ" = "1" ] && {
LOCAL_DNS="127.0.0.1#${NEXT_DNS_LISTEN_PORT}"
local china_ng_c_dns="tcp://$(get_first_dns DIRECT_DNS 53 | sed 's/:/#/g')"
ln_run "$(first_type chinadns-ng)" chinadns-ng "/dev/null" -b :: -l ${NEXT_DNS_LISTEN_PORT} -c ${china_ng_c_dns} -d chn
echolog " - ChinaDNS-NG(${LOCAL_DNS}) -> ${china_ng_c_dns}"
echolog " * 请确保上游直连 DNS 支持 TCP 查询。"
NEXT_DNS_LISTEN_PORT=$(expr $NEXT_DNS_LISTEN_PORT + 1)
}
;;
auto)
#Automatic logic is already done by default
:
;;
esac
# 追加直连DNS到iptables/nftables
[ "$(config_t_get global_haproxy balancing_enable 0)" != "1" ] && IPT_APPEND_DNS=
add_default_port() {
[ -z "$1" ] && echo "" || echo "$1" | awk -F',' '{for(i=1;i<=NF;i++){if($i !~ /#/) $i=$i"#53";} print $0;}' OFS=','
}
LOCAL_DNS=$(add_default_port "$LOCAL_DNS")
IPT_APPEND_DNS=$(add_default_port "${IPT_APPEND_DNS:-$LOCAL_DNS}")
echo "$IPT_APPEND_DNS" | grep -q -E "(^|,)$LOCAL_DNS(,|$)" || IPT_APPEND_DNS="${IPT_APPEND_DNS:+$IPT_APPEND_DNS,}$LOCAL_DNS"
[ -n "$DIRECT_DNS" ] && {
DIRECT_DNS=$(add_default_port "$DIRECT_DNS")
echo "$IPT_APPEND_DNS" | grep -q -E "(^|,)$DIRECT_DNS(,|$)" || IPT_APPEND_DNS="${IPT_APPEND_DNS:+$IPT_APPEND_DNS,}$DIRECT_DNS"
}
# 排除127.0.0.1的条目
IPT_APPEND_DNS=$(echo "$IPT_APPEND_DNS" | awk -F',' '{for(i=1;i<=NF;i++) if($i !~ /^127\.0\.0\.1/) printf (i>1?",":"") $i; print ""}' | sed 's/^,\|,$//g')
TUN_DNS="127.0.0.1#${NEXT_DNS_LISTEN_PORT}"
[ -n "${NO_PLUGIN_DNS}" ] && TUN_DNS="127.0.0.1#${resolve_dns_port}"
case "$DNS_MODE" in
dns2socks)
local dns2socks_socks_server=$(echo $(config_t_get global socks_server 127.0.0.1:1080) | sed "s/#/:/g")
local dns2socks_forward=$(get_first_dns REMOTE_DNS 53 | sed 's/#/:/g')
run_dns2socks socks=$dns2socks_socks_server listen_address=127.0.0.1 listen_port=${NEXT_DNS_LISTEN_PORT} dns=$dns2socks_forward cache=$DNS_CACHE
echolog " - dns2socks(${TUN_DNS}),${dns2socks_socks_server} -> tcp://${dns2socks_forward}"
;;
sing-box)
[ -z "${NO_PLUGIN_DNS}" ] && {
local config_file=$TMP_PATH/DNS.json
local log_file=$TMP_PATH/DNS.log
local log_file=/dev/null
local _args="type=$DNS_MODE config_file=$config_file log_file=$log_file"
[ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0"
_args="${_args} direct_dns_query_strategy=${DIRECT_DNS_QUERY_STRATEGY}"
_args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}"
DNSMASQ_FILTER_PROXY_IPV6=0
local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip)
[ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}"
TCP_PROXY_DNS=1
local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp)
_args="${_args} dns_listen_port=${NEXT_DNS_LISTEN_PORT}"
_args="${_args} remote_dns_protocol=${v2ray_dns_mode}"
case "$v2ray_dns_mode" in
tcp)
_args="${_args} remote_dns_tcp_server=${REMOTE_DNS}"
echolog " - Sing-Box DNS(${TUN_DNS}) -> tcp://${REMOTE_DNS}"
;;
doh)
remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query")
_args="${_args} remote_dns_doh=${remote_dns_doh}"
echolog " - Sing-Box DNS(${TUN_DNS}) -> ${remote_dns_doh}"
local _doh_url _doh_host _doh_port _doh_bootstrap
parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap
[ -n "${_doh_bootstrap}" ] && REMOTE_DNS="${_doh_bootstrap}#${_doh_port}"
;;
esac
_args="${_args} dns_socks_address=127.0.0.1 dns_socks_port=${tcp_node_socks_port}"
[ -n "${sing_box_local_dns}" ] && _args="${_args} ${sing_box_local_dns}"
run_singbox ${_args}
}
;;
xray)
[ -z "${NO_PLUGIN_DNS}" ] && {
local config_file=$TMP_PATH/DNS.json
local log_file=$TMP_PATH/DNS.log
local log_file=/dev/null
local _args="type=$DNS_MODE config_file=$config_file log_file=$log_file"
[ "${DNS_CACHE}" == "0" ] && _args="${_args} dns_cache=0"
_args="${_args} direct_dns_query_strategy=${DIRECT_DNS_QUERY_STRATEGY}"
_args="${_args} remote_dns_query_strategy=${REMOTE_DNS_QUERY_STRATEGY}"
DNSMASQ_FILTER_PROXY_IPV6=0
local _remote_dns_client_ip=$(config_t_get global remote_dns_client_ip)
[ -n "${_remote_dns_client_ip}" ] && _args="${_args} remote_dns_client_ip=${_remote_dns_client_ip}"
TCP_PROXY_DNS=1
_args="${_args} dns_listen_port=${NEXT_DNS_LISTEN_PORT}"
_args="${_args} remote_dns_tcp_server=${REMOTE_DNS}"
local v2ray_dns_mode=$(config_t_get global v2ray_dns_mode tcp)
if [ "$v2ray_dns_mode" = "tcp+doh" ]; then
remote_dns_doh=$(config_t_get global remote_dns_doh "https://1.1.1.1/dns-query")
_args="${_args} remote_dns_doh=${remote_dns_doh}"
echolog " - Xray DNS(${TUN_DNS}) -> (${remote_dns_doh})(A/AAAA) + tcp://${REMOTE_DNS}"
local _doh_url _doh_host _doh_port _doh_bootstrap
parse_doh "$remote_dns_doh" _doh_url _doh_host _doh_port _doh_bootstrap
[ -n "${_doh_bootstrap}" ] && REMOTE_DNS="${REMOTE_DNS},${_doh_bootstrap}#${_doh_port}"
else
echolog " - Xray DNS(${TUN_DNS}) -> tcp://${REMOTE_DNS}"
fi
_args="${_args} dns_socks_address=127.0.0.1 dns_socks_port=${tcp_node_socks_port}"
run_xray ${_args}
}
;;
udp)
UDP_PROXY_DNS=1
local china_ng_listen_port=${NEXT_DNS_LISTEN_PORT}
local china_ng_trust_dns="udp://$(get_first_dns REMOTE_DNS 53 | sed 's/:/#/g')"
if [ "$DNS_SHUNT" != "chinadns-ng" ] && [ "$FILTER_PROXY_IPV6" = "1" ]; then
DNSMASQ_FILTER_PROXY_IPV6=0
local no_ipv6_trust="-N"
ln_run "$(first_type chinadns-ng)" chinadns-ng "/dev/null" -b :: -l ${china_ng_listen_port} -t ${china_ng_trust_dns} -d gfw ${no_ipv6_trust}
echolog " - ChinaDNS-NG(${TUN_DNS}) -> ${china_ng_trust_dns}"
else
TUN_DNS="$(echo ${REMOTE_DNS} | sed 's/#/:/g' | sed -E 's/\:([^:]+)$/#\1/g')"
echolog " - udp://${TUN_DNS}"
fi
;;
tcp)
TCP_PROXY_DNS=1
local china_ng_listen_port=${NEXT_DNS_LISTEN_PORT}
local china_ng_trust_dns="tcp://$(get_first_dns REMOTE_DNS 53 | sed 's/:/#/g')"
[ "$DNS_SHUNT" != "chinadns-ng" ] && {
[ "$FILTER_PROXY_IPV6" = "1" ] && DNSMASQ_FILTER_PROXY_IPV6=0 && local no_ipv6_trust="-N"
ln_run "$(first_type chinadns-ng)" chinadns-ng "/dev/null" -b :: -l ${china_ng_listen_port} -t ${china_ng_trust_dns} -d gfw ${no_ipv6_trust}
echolog " - ChinaDNS-NG(${TUN_DNS}) -> ${china_ng_trust_dns}"
}
;;
esac
[ -n "${resolve_dns_log}" ] && echolog " - ${resolve_dns_log}"
[ -n "${TCP_PROXY_DNS}" ] && echolog " * 请确认上游 DNS 支持 TCP/DoH 查询,如非直连地址,确保 TCP 代理打开,并且已经正确转发!"
[ -n "${UDP_PROXY_DNS}" ] && echolog " * 请确认上游 DNS 支持 UDP 查询并已使用 UDP 节点,如上游 DNS 非直连地址,确保 UDP 代理打开,并且已经正确转发!"
[ "${DNS_SHUNT}" = "smartdns" ] && {
if command -v smartdns > /dev/null 2>&1; then
rm -rf $TMP_PATH2/dnsmasq_default*
local group_domestic=$(config_t_get global group_domestic)
local smartdns_remote_dns=$(config_t_get global smartdns_remote_dns)
if [ -n "${smartdns_remote_dns}" -a "${smartdns_remote_dns}" != "nil" ]; then
smartdns_remote_dns=$(echo ${smartdns_remote_dns} | tr -s ' ' '|')
else
smartdns_remote_dns="tcp://1.1.1.1"
fi
local subnet_ip=$(config_t_get global remote_dns_client_ip)
lua $APP_PATH/helper_smartdns_add.lua -FLAG "default" -SMARTDNS_CONF "/tmp/etc/smartdns/$CONFIG.conf" \
-LOCAL_GROUP ${group_domestic:-nil} -REMOTE_GROUP "passwall_proxy" -REMOTE_PROXY_SERVER ${TCP_SOCKS_server} -USE_DEFAULT_DNS "${USE_DEFAULT_DNS:-direct}" \
-REMOTE_DNS ${smartdns_remote_dns} -DNS_MODE ${DNS_MODE:-socks} -TUN_DNS ${TUN_DNS} -REMOTE_FAKEDNS ${fakedns:-0} \
-USE_DIRECT_LIST "${USE_DIRECT_LIST}" -USE_PROXY_LIST "${USE_PROXY_LIST}" -USE_BLOCK_LIST "${USE_BLOCK_LIST}" -USE_GFW_LIST "${USE_GFW_LIST}" -CHN_LIST "${CHN_LIST}" \
-TCP_NODE ${TCP_NODE} -DEFAULT_PROXY_MODE "${TCP_PROXY_MODE}" -NO_PROXY_IPV6 ${FILTER_PROXY_IPV6:-0} -NFTFLAG ${nftflag:-0} \
-SUBNET ${subnet_ip:-0} -NO_LOGIC_LOG ${NO_LOGIC_LOG:-0}
source $APP_PATH/helper_smartdns.sh restart
echolog " - 域名解析:使用SmartDNS,请确保配置正常。"
return
else
DNS_SHUNT="dnsmasq"
echolog " * 未安装SmartDNS,默认使用Dnsmasq进行域名解析!"
fi
}
[ "$DNS_SHUNT" = "chinadns-ng" ] && [ -n "$(first_type chinadns-ng)" ] && {
chinadns_ng_min=2024.04.13
chinadns_ng_now=$($(first_type chinadns-ng) -V | grep -i "ChinaDNS-NG " | awk '{print $2}')
if [ $(check_ver "$chinadns_ng_now" "$chinadns_ng_min") = 1 ]; then
echolog " * 注意:当前 ChinaDNS-NG 版本为[ $chinadns_ng_now ],请更新到[ $chinadns_ng_min ]或以上版本,否则 DNS 有可能无法正常工作!"
fi
[ "$FILTER_PROXY_IPV6" = "1" ] && DNSMASQ_FILTER_PROXY_IPV6=0
[ -z "${china_ng_listen_port}" ] && local china_ng_listen_port=$(expr $NEXT_DNS_LISTEN_PORT + 1)
local china_ng_listen="127.0.0.1#${china_ng_listen_port}"
[ -z "${china_ng_trust_dns}" ] && local china_ng_trust_dns=${TUN_DNS}
echolog " - ChinaDNS-NG(${china_ng_listen}):直连DNS:${china_ng_local_dns},可信DNS:${china_ng_trust_dns}"
china_ng_listen="${china_ng_listen},::1#${china_ng_listen_port}"
run_chinadns_ng \
_flag="default" \
_listen_port=${china_ng_listen_port} \
_dns_local=${china_ng_local_dns} \
_dns_trust=${china_ng_trust_dns} \
_no_ipv6_trust=${FILTER_PROXY_IPV6} \
_use_direct_list=${USE_DIRECT_LIST} \
_use_proxy_list=${USE_PROXY_LIST} \
_use_block_list=${USE_BLOCK_LIST} \
_gfwlist=${USE_GFW_LIST} \
_chnlist=${CHN_LIST} \
_default_mode=${TCP_PROXY_MODE} \
_default_tag=$(config_t_get global chinadns_ng_default_tag smart) \
_no_logic_log=0 \
_tcp_node=${TCP_NODE} \
_remote_fakedns=${fakedns:-0}
USE_DEFAULT_DNS="chinadns_ng"
}
[ "$USE_DEFAULT_DNS" = "remote" ] && {
dnsmasq_version=$(dnsmasq -v | grep -i "Dnsmasq version " | awk '{print $3}')
[ "$(expr $dnsmasq_version \>= 2.87)" == 0 ] && echolog "Dnsmasq版本低于2.87,有可能无法正常使用!!!"
}
local DNSMASQ_TUN_DNS=$(get_first_dns TUN_DNS 53 | sed 's/:/#/g')
local RUN_NEW_DNSMASQ=1
RUN_NEW_DNSMASQ=${DNS_REDIRECT}
if [ "${RUN_NEW_DNSMASQ}" == "0" ]; then
#The old logic will be removed in the future.
#Run a copy dnsmasq instance, DNS hijack that don't need a proxy devices.
[ "1" = "0" ] && {
DIRECT_DNSMASQ_PORT=$(get_new_port 11400)
DIRECT_DNSMASQ_CONF=${GLOBAL_ACL_PATH}/direct_dnsmasq.conf
DIRECT_DNSMASQ_CONF_PATH=${GLOBAL_ACL_PATH}/direct_dnsmasq.d
mkdir -p ${DIRECT_DNSMASQ_CONF_PATH}
lua $APP_PATH/helper_dnsmasq.lua copy_instance -LISTEN_PORT ${DIRECT_DNSMASQ_PORT} -DNSMASQ_CONF ${DIRECT_DNSMASQ_CONF} -TMP_DNSMASQ_PATH ${DIRECT_DNSMASQ_CONF_PATH}
ln_run "$(first_type dnsmasq)" "dnsmasq_direct" "/dev/null" -C ${DIRECT_DNSMASQ_CONF} -x ${GLOBAL_ACL_PATH}/direct_dnsmasq.pid
echo "${DIRECT_DNSMASQ_PORT}" > ${GLOBAL_ACL_PATH}/direct_dnsmasq_port
}
#Rewrite the default DNS service configuration
#Modify the default dnsmasq service
lua $APP_PATH/helper_dnsmasq.lua stretch
lua $APP_PATH/helper_dnsmasq.lua add_rule -FLAG "default" -TMP_DNSMASQ_PATH ${GLOBAL_DNSMASQ_CONF_PATH} -DNSMASQ_CONF_FILE ${GLOBAL_DNSMASQ_CONF} \
-DEFAULT_DNS ${DEFAULT_DNS} -LOCAL_DNS ${LOCAL_DNS} -TUN_DNS ${DNSMASQ_TUN_DNS} \
-REMOTE_FAKEDNS ${fakedns:-0} -USE_DEFAULT_DNS "${USE_DEFAULT_DNS:-direct}" -CHINADNS_DNS ${china_ng_listen:-0} \
-USE_DIRECT_LIST "${USE_DIRECT_LIST}" -USE_PROXY_LIST "${USE_PROXY_LIST}" -USE_BLOCK_LIST "${USE_BLOCK_LIST}" -USE_GFW_LIST "${USE_GFW_LIST}" -CHN_LIST "${CHN_LIST}" \
-TCP_NODE ${TCP_NODE} -DEFAULT_PROXY_MODE ${TCP_PROXY_MODE} -NO_PROXY_IPV6 ${DNSMASQ_FILTER_PROXY_IPV6:-0} -NFTFLAG ${nftflag:-0} \
-NO_LOGIC_LOG ${NO_LOGIC_LOG:-0}
uci -q add_list dhcp.@dnsmasq[0].addnmount=${GLOBAL_DNSMASQ_CONF_PATH}
uci -q commit dhcp
lua $APP_PATH/helper_dnsmasq.lua logic_restart -LOG 1
else
#Run a copy dnsmasq instance, DNS hijack for that need proxy devices.
GLOBAL_DNSMASQ_PORT=$(get_new_port 11400)
GLOBAL_DNSMASQ_CONF=${GLOBAL_ACL_PATH}/dnsmasq.conf
GLOBAL_DNSMASQ_CONF_PATH=${GLOBAL_ACL_PATH}/dnsmasq.d
lua $APP_PATH/helper_dnsmasq.lua add_rule -FLAG "default" -TMP_DNSMASQ_PATH ${GLOBAL_DNSMASQ_CONF_PATH} -DNSMASQ_CONF_FILE ${GLOBAL_DNSMASQ_CONF} \
-LISTEN_PORT ${GLOBAL_DNSMASQ_PORT} -DEFAULT_DNS ${DEFAULT_DNS} -LOCAL_DNS ${LOCAL_DNS} -TUN_DNS ${DNSMASQ_TUN_DNS} \
-REMOTE_FAKEDNS ${fakedns:-0} -USE_DEFAULT_DNS "${USE_DEFAULT_DNS:-direct}" -CHINADNS_DNS ${china_ng_listen:-0} \
-USE_DIRECT_LIST "${USE_DIRECT_LIST}" -USE_PROXY_LIST "${USE_PROXY_LIST}" -USE_BLOCK_LIST "${USE_BLOCK_LIST}" -USE_GFW_LIST "${USE_GFW_LIST}" -CHN_LIST "${CHN_LIST}" \
-TCP_NODE ${TCP_NODE} -DEFAULT_PROXY_MODE ${TCP_PROXY_MODE} -NO_PROXY_IPV6 ${DNSMASQ_FILTER_PROXY_IPV6:-0} -NFTFLAG ${nftflag:-0} \
-NO_LOGIC_LOG ${NO_LOGIC_LOG:-0}
ln_run "$(first_type dnsmasq)" "dnsmasq_default" "/dev/null" -C ${GLOBAL_DNSMASQ_CONF} -x ${GLOBAL_ACL_PATH}/dnsmasq.pid
set_cache_var "ACL_default_dns_port" "${GLOBAL_DNSMASQ_PORT}"
DNS_REDIRECT_PORT=${GLOBAL_DNSMASQ_PORT}
#dhcp.leases to hosts
$APP_PATH/lease2hosts.sh > /dev/null 2>&1 &
fi
}
add_ip2route() {
local ip=$(get_host_ip "ipv4" $1)
[ -z "$ip" ] && {
echolog " - 无法解析[${1}],路由表添加失败!"
return 1
}
local remarks="${1}"
[ "$remarks" != "$ip" ] && remarks="${1}(${ip})"
. /lib/functions/network.sh
local gateway device
network_get_gateway gateway "$2"
network_get_device device "$2"
[ -z "${device}" ] && device="$2"
if [ -n "${gateway}" ]; then
route add -host ${ip} gw ${gateway} dev ${device} >/dev/null 2>&1
echo "$ip" >> $TMP_ROUTE_PATH/${device}
echolog " - [${remarks}]添加到接口[${device}]路由表成功!"
else
echolog " - [${remarks}]添加到接口[${device}]路由表失功!原因是找不到[${device}]网关。"
fi
}
delete_ip2route() {
[ -d "${TMP_ROUTE_PATH}" ] && {
local interface
for interface in $(ls ${TMP_ROUTE_PATH}); do
for ip in $(cat ${TMP_ROUTE_PATH}/${interface}); do
route del -host ${ip} dev ${interface} >/dev/null 2>&1
done
done
}
}
start_haproxy() {
[ "$(config_t_get global_haproxy balancing_enable 0)" != "1" ] && return
haproxy_path=$TMP_PATH/haproxy
haproxy_conf="config.cfg"
lua $APP_PATH/haproxy.lua -path ${haproxy_path} -conf ${haproxy_conf} -dns ${LOCAL_DNS}
ln_run "$(first_type haproxy)" haproxy "/dev/null" -f "${haproxy_path}/${haproxy_conf}"
}
kill_all() {
kill -9 $(pidof "$@") >/dev/null 2>&1
}
acl_app() {
local items=$(uci show ${CONFIG} | grep "=acl_rule" | cut -d '.' -sf 2 | cut -d '=' -sf 1)
[ -n "$items" ] && {
local item
local socks_port redir_port dns_port dnsmasq_port chinadns_port
local msg msg2
socks_port=11100
redir_port=11200
dns_port=11300
dnsmasq_port=${GLOBAL_DNSMASQ_PORT:-11400}
chinadns_port=11500
for item in $items; do
local sid=$(uci -q show "${CONFIG}.${item}" | grep "=acl_rule" | awk -F '=' '{print $1}' | awk -F '.' '{print $2}')
[ "$(config_n_get $sid enabled)" = "1" ] || continue
eval $(uci -q show "${CONFIG}.${item}" | cut -d'.' -sf 3-)
if [ -n "${sources}" ]; then
for s in $sources; do
local s2
is_iprange=$(lua_api "iprange(\"${s}\")")
if [ "${is_iprange}" = "true" ]; then
s2="iprange:${s}"
elif [ -n "$(echo ${s} | grep '^ipset:')" ]; then
s2="ipset:${s}"
else
_ip_or_mac=$(lua_api "ip_or_mac(\"${s}\")")
if [ "${_ip_or_mac}" = "ip" ]; then
s2="ip:${s}"
elif [ "${_ip_or_mac}" = "mac" ]; then
s2="mac:${s}"
fi
fi
[ -n "${s2}" ] && source_list="${source_list}\n${s2}"
unset s2
done
else
source_list="any"
fi
local acl_path=${TMP_ACL_PATH}/$sid
mkdir -p ${acl_path}
[ -n "${source_list}" ] && echo -e "${source_list}" | sed '/^$/d' > ${acl_path}/source_list
use_global_config=${use_global_config:-0}
[ "${use_global_config}" = "1" ] && {
tcp_node="default"
udp_node="default"
}
tcp_no_redir_ports=${tcp_no_redir_ports:-${TCP_NO_REDIR_PORTS}}
udp_no_redir_ports=${udp_no_redir_ports:-${UDP_NO_REDIR_PORTS}}
if has_1_65535 "$tcp_no_redir_ports" && has_1_65535 "$udp_no_redir_ports"; then
unset use_global_config
unset tcp_node
unset udp_node
else
use_direct_list=${use_direct_list:-1}
use_proxy_list=${use_proxy_list:-1}
use_block_list=${use_block_list:-1}
use_gfw_list=${use_gfw_list:-1}
chn_list=${chn_list:-direct}
tcp_proxy_mode=${tcp_proxy_mode:-proxy}
udp_proxy_mode=${udp_proxy_mode:-proxy}
filter_proxy_ipv6=${filter_proxy_ipv6:-0}
dnsmasq_filter_proxy_ipv6=${filter_proxy_ipv6}
dns_shunt=${dns_shunt:-dnsmasq}
dns_mode=${dns_mode:-dns2socks}
remote_dns=${remote_dns:-1.1.1.1}
use_default_dns=${use_default_dns:-direct}
[ "$dns_mode" = "sing-box" ] && {
[ "$v2ray_dns_mode" = "doh" ] && remote_dns=${remote_dns_doh:-https://1.1.1.1/dns-query}
}
fi
[ -n "$tcp_node" ] && {
local GLOBAL_TCP_NODE=$(get_cache_var "ACL_GLOBAL_TCP_node")
[ -n "${GLOBAL_TCP_NODE}" ] && GLOBAL_TCP_redir_port=$(get_cache_var "ACL_GLOBAL_TCP_redir_port")
if [ "$tcp_node" = "default" ]; then
if [ -n "${GLOBAL_TCP_NODE}" ]; then
set_cache_var "ACL_${sid}_tcp_node" "${GLOBAL_TCP_NODE}"
set_cache_var "ACL_${sid}_tcp_redir_port" "${GLOBAL_TCP_redir_port}"
set_cache_var "ACL_${sid}_dns_port" "${GLOBAL_DNSMASQ_PORT}"
set_cache_var "ACL_${sid}_tcp_default" "1"
else
echolog " - 全局节点未启用,跳过【${remarks}】"
fi
else
[ "$(config_get_type $tcp_node)" = "nodes" ] && {
if [ -n "${GLOBAL_TCP_NODE}" ] && [ "$tcp_node" = "${GLOBAL_TCP_NODE}" ]; then
set_cache_var "ACL_${sid}_tcp_node" "${GLOBAL_TCP_NODE}"
set_cache_var "ACL_${sid}_tcp_redir_port" "${GLOBAL_TCP_redir_port}"
set_cache_var "ACL_${sid}_dns_port" "${GLOBAL_DNSMASQ_PORT}"
set_cache_var "ACL_${sid}_tcp_default" "1"
else
run_dns() {
local _dns_port
[ -n $1 ] && _dns_port=$1
[ -z ${_dns_port} ] && {
dns_port=$(get_new_port $(expr $dns_port + 1))
_dns_port=$dns_port
if [ "$dns_mode" = "dns2socks" ]; then
run_dns2socks flag=acl_${sid} socks_address=127.0.0.1 socks_port=$socks_port listen_address=0.0.0.0 listen_port=${_dns_port} dns=$remote_dns cache=1
elif [ "$dns_mode" = "sing-box" -o "$dns_mode" = "xray" ]; then
config_file=$TMP_ACL_PATH/${tcp_node}_SOCKS_${socks_port}_DNS.json
[ "$dns_mode" = "xray" ] && [ "$v2ray_dns_mode" = "tcp+doh" ] && remote_dns_doh=${remote_dns_doh:-https://1.1.1.1/dns-query}
local type=${dns_mode}
[ "${dns_mode}" = "sing-box" ] && type="singbox"
dnsmasq_filter_proxy_ipv6=0
remote_dns_query_strategy="UseIP"
[ "$filter_proxy_ipv6" = "1" ] && remote_dns_query_strategy="UseIPv4"
run_${type} flag=acl_${sid} type=$dns_mode dns_socks_address=127.0.0.1 dns_socks_port=$socks_port dns_listen_port=${_dns_port} remote_dns_protocol=${v2ray_dns_mode} remote_dns_tcp_server=${remote_dns} remote_dns_doh="${remote_dns_doh}" remote_dns_query_strategy=${remote_dns_query_strategy} remote_dns_client_ip=${remote_dns_client_ip} config_file=$config_file
fi
set_cache_var "node_${tcp_node}_$(echo -n "${remote_dns}" | md5sum | cut -d " " -f1)" "${_dns_port}"
}
[ "$dns_shunt" = "chinadns-ng" ] && [ -n "$(first_type chinadns-ng)" ] && {
chinadns_ng_min=2024.04.13
chinadns_ng_now=$($(first_type chinadns-ng) -V | grep -i "ChinaDNS-NG " | awk '{print $2}')
if [ $(check_ver "$chinadns_ng_now" "$chinadns_ng_min") = 1 ]; then
echolog " * 注意:当前 ChinaDNS-NG 版本为[ $chinadns_ng_now ],请更新到[ $chinadns_ng_min ]或以上版本,否则 DNS 有可能无法正常工作!"
fi
[ "$filter_proxy_ipv6" = "1" ] && dnsmasq_filter_proxy_ipv6=0
chinadns_port=$(expr $chinadns_port + 1)
_china_ng_listen="127.0.0.1#${chinadns_port},::1#${chinadns_port}"
_chinadns_local_dns=$(IFS=','; set -- $LOCAL_DNS; [ "${1%%[#:]*}" = "127.0.0.1" ] && echo "$1" || ([ -n "$2" ] && echo "$1,$2" || echo "$1"))
_direct_dns_mode=$(config_t_get global direct_dns_mode "auto")
case "${_direct_dns_mode}" in
udp)
_chinadns_local_dns=$(config_t_get global direct_dns_udp 223.5.5.5 | sed 's/:/#/g')
;;
tcp)
_chinadns_local_dns="tcp://$(config_t_get global direct_dns_tcp 223.5.5.5 | sed 's/:/#/g')"
;;
esac
run_chinadns_ng \
_flag="$sid" \
_listen_port=${chinadns_port} \
_dns_local=${_chinadns_local_dns} \
_dns_trust=127.0.0.1#${_dns_port} \
_no_ipv6_trust=${filter_proxy_ipv6} \
_use_direct_list=${use_direct_list} \
_use_proxy_list=${use_proxy_list} \
_use_block_list=${use_block_list} \
_gfwlist=${use_gfw_list} \
_chnlist=${chn_list} \
_default_mode=${tcp_proxy_mode} \
_default_tag=${chinadns_ng_default_tag:-smart} \
_no_logic_log=1 \
_tcp_node=${tcp_node} \
_remote_fakedns=0
use_default_dns="chinadns_ng"
}
dnsmasq_port=$(get_new_port $(expr $dnsmasq_port + 1))
local dnsmasq_conf=${acl_path}/dnsmasq.conf
local dnsmasq_conf_path=${acl_path}/dnsmasq.d
lua $APP_PATH/helper_dnsmasq.lua add_rule -FLAG ${sid} -TMP_DNSMASQ_PATH ${dnsmasq_conf_path} -DNSMASQ_CONF_FILE ${dnsmasq_conf} \
-LISTEN_PORT ${dnsmasq_port} -DEFAULT_DNS ${DEFAULT_DNS} -LOCAL_DNS $LOCAL_DNS \
-USE_DIRECT_LIST "${use_direct_list}" -USE_PROXY_LIST "${use_proxy_list}" -USE_BLOCK_LIST "${use_block_list}" -USE_GFW_LIST "${use_gfw_list}" -CHN_LIST "${chn_list}" \
-TUN_DNS "127.0.0.1#${_dns_port}" -REMOTE_FAKEDNS 0 -USE_DEFAULT_DNS "${use_default_dns:-direct}" -CHINADNS_DNS ${_china_ng_listen:-0} \
-TCP_NODE $tcp_node -DEFAULT_PROXY_MODE ${tcp_proxy_mode} -NO_PROXY_IPV6 ${dnsmasq_filter_proxy_ipv6:-0} -NFTFLAG ${nftflag:-0} \
-NO_LOGIC_LOG 1
ln_run "$(first_type dnsmasq)" "dnsmasq_${sid}" "/dev/null" -C ${dnsmasq_conf} -x ${acl_path}/dnsmasq.pid
set_cache_var "ACL_${sid}_dns_port" "${dnsmasq_port}"
set_cache_var "node_${tcp_node}_$(echo -n "${tcp_proxy_mode}${remote_dns}" | md5sum | cut -d " " -f1)" "${dnsmasq_port}"
#dhcp.leases to hosts
$APP_PATH/lease2hosts.sh > /dev/null 2>&1 &
}
_redir_port=$(get_cache_var "node_${tcp_node}_redir_port")
_socks_port=$(get_cache_var "node_${tcp_node}_socks_port")
if [ -n "${_socks_port}" ] && [ -n "${_redir_port}" ]; then
socks_port=${_socks_port}
tcp_port=${_redir_port}
_dnsmasq_port=$(get_cache_var "node_${tcp_node}_$(echo -n "${tcp_proxy_mode}${remote_dns}" | md5sum | cut -d " " -f1)")
if [ -z "${_dnsmasq_port}" ]; then
_dns_port=$(get_cache_var "node_${tcp_node}_$(echo -n "${remote_dns}" | md5sum | cut -d " " -f1)")
run_dns ${_dns_port}
else
[ -n "${_dnsmasq_port}" ] && set_cache_var "ACL_${sid}_dns_port" "${_dnsmasq_port}"
fi
else
socks_port=$(get_new_port $(expr $socks_port + 1))
set_cache_var "node_${tcp_node}_socks_port" "${socks_port}"
redir_port=$(get_new_port $(expr $redir_port + 1))
set_cache_var "node_${tcp_node}_redir_port" "${redir_port}"
tcp_port=$redir_port
local type=$(echo $(config_n_get $tcp_node type) | tr 'A-Z' 'a-z')
if [ -n "${type}" ] && ([ "${type}" = "sing-box" ] || [ "${type}" = "xray" ]); then
config_file="acl/${tcp_node}_TCP_${redir_port}.json"
_extra_param="socks_address=127.0.0.1 socks_port=$socks_port"
_extra_param="${_extra_param} tcp_proxy_way=$TCP_PROXY_WAY"
if [ "$dns_mode" = "sing-box" ] || [ "$dns_mode" = "xray" ]; then
dns_port=$(get_new_port $(expr $dns_port + 1))
_dns_port=$dns_port
config_file="${config_file//TCP_/DNS_${_dns_port}_TCP_}"
remote_dns_doh=${remote_dns}
dnsmasq_filter_proxy_ipv6=0
remote_dns_query_strategy="UseIP"
[ "$filter_proxy_ipv6" = "1" ] && remote_dns_query_strategy="UseIPv4"
[ "$dns_mode" = "xray" ] && [ "$v2ray_dns_mode" = "tcp+doh" ] && remote_dns_doh=${remote_dns_doh:-https://1.1.1.1/dns-query}
_extra_param="dns_listen_port=${_dns_port} remote_dns_protocol=${v2ray_dns_mode} remote_dns_tcp_server=${remote_dns} remote_dns_doh=${remote_dns_doh} remote_dns_query_strategy=${remote_dns_query_strategy} remote_dns_client_ip=${remote_dns_client_ip}"
fi
[ -n "$udp_node" ] && ([ "$udp_node" = "tcp" ] || [ "$udp_node" = "$tcp_node" ]) && {
config_file="${config_file//TCP_/TCP_UDP_}"
_extra_param="${_extra_param} udp_redir_port=$redir_port"
}
config_file="$TMP_PATH/$config_file"
[ "${type}" = "sing-box" ] && type="singbox"
run_${type} flag=$tcp_node node=$tcp_node tcp_redir_port=$redir_port ${_extra_param} config_file=$config_file
else
config_file="acl/${tcp_node}_SOCKS_${socks_port}.json"
run_socks flag=$tcp_node node=$tcp_node bind=127.0.0.1 socks_port=$socks_port config_file=$config_file
local log_file=$TMP_ACL_PATH/ipt2socks_${tcp_node}_${redir_port}.log
log_file="/dev/null"
run_ipt2socks flag=acl_${tcp_node} tcp_tproxy=${is_tproxy} local_port=$redir_port socks_address=127.0.0.1 socks_port=$socks_port log_file=$log_file
fi
run_dns ${_dns_port}
fi
set_cache_var "ACL_${sid}_tcp_node" "${tcp_node}"
set_cache_var "ACL_${sid}_tcp_redir_port" "${tcp_port}"
fi
}
fi
}
[ -n "$udp_node" ] && {
if [ "$udp_node" = "default" ]; then
local GLOBAL_UDP_NODE=$(get_cache_var "ACL_GLOBAL_UDP_node")
[ -n "${GLOBAL_UDP_NODE}" ] && GLOBAL_UDP_redir_port=$(get_cache_var "ACL_GLOBAL_UDP_redir_port")
if [ -n "${GLOBAL_UDP_NODE}" ]; then
set_cache_var "ACL_${sid}_udp_node" "${GLOBAL_UDP_NODE}"
set_cache_var "ACL_${sid}_udp_redir_port" "${GLOBAL_UDP_redir_port}"
set_cache_var "ACL_${sid}_udp_default" "1"
else
echolog " - 全局节点未启用,跳过【${remarks}】"
fi
elif [ "$udp_node" = "tcp" ] || [ "$udp_node" = "$tcp_node" ]; then
udp_node=$(get_cache_var "ACL_${sid}_tcp_node")
udp_port=$(get_cache_var "ACL_${sid}_tcp_redir_port")
set_cache_var "ACL_${sid}_udp_node" "${udp_node}"
set_cache_var "ACL_${sid}_udp_redir_port" "${udp_port}"
else
[ "$(config_get_type $udp_node)" = "nodes" ] && {
if [ -n "${GLOBAL_UDP_NODE}" ] && [ "$udp_node" = "${GLOBAL_UDP_NODE}" ]; then
set_cache_var "ACL_${sid}_udp_node" "${GLOBAL_UDP_NODE}"
set_cache_var "ACL_${sid}_udp_redir_port" "${GLOBAL_UDP_redir_port}"
set_cache_var "ACL_${sid}_udp_default" "1"
else
_redir_port=$(get_cache_var "node_${udp_node}_redir_port")
_socks_port=$(get_cache_var "node_${udp_node}_socks_port")
if [ -n "${_socks_port}" ] && [ -n "${_redir_port}" ]; then
socks_port=${_socks_port}
udp_port=${_redir_port}
else
socks_port=$(get_new_port $(expr $socks_port + 1))
set_cache_var "node_${udp_node}_socks_port" "${socks_port}"
redir_port=$(get_new_port $(expr $redir_port + 1))
set_cache_var "node_${udp_node}_redir_port" "${redir_port}"
udp_port=$redir_port
local type=$(echo $(config_n_get $udp_node type) | tr 'A-Z' 'a-z')
if [ -n "${type}" ] && ([ "${type}" = "sing-box" ] || [ "${type}" = "xray" ]); then
config_file="acl/${udp_node}_UDP_${redir_port}.json"
config_file="$TMP_PATH/$config_file"
[ "${type}" = "sing-box" ] && type="singbox"
run_${type} flag=$udp_node node=$udp_node udp_redir_port=$redir_port config_file=$config_file
else
config_file="acl/${udp_node}_SOCKS_${socks_port}.json"
run_socks flag=$udp_node node=$udp_node bind=127.0.0.1 socks_port=$socks_port config_file=$config_file
local log_file=$TMP_ACL_PATH/ipt2socks_${udp_node}_${redir_port}.log
log_file="/dev/null"
run_ipt2socks flag=acl_${udp_node} local_port=$redir_port socks_address=127.0.0.1 socks_port=$socks_port log_file=$log_file
fi
fi
set_cache_var "ACL_${sid}_udp_node" "${udp_node}"
set_cache_var "ACL_${sid}_udp_redir_port" "${udp_port}"
fi
}
fi
}
unset enabled sid remarks sources interface tcp_no_redir_ports udp_no_redir_ports use_global_config tcp_node udp_node use_direct_list use_proxy_list use_block_list use_gfw_list chn_list tcp_proxy_mode udp_proxy_mode filter_proxy_ipv6 dns_mode remote_dns v2ray_dns_mode remote_dns_doh remote_dns_client_ip
unset _ip _mac _iprange _ipset _ip_or_mac source_list tcp_port udp_port config_file _extra_param
unset _china_ng_listen _chinadns_local_dns _direct_dns_mode chinadns_ng_default_tag dnsmasq_filter_proxy_ipv6
done
unset socks_port redir_port dns_port dnsmasq_port chinadns_port
}
}
start() {
mkdir -p /tmp/etc /tmp/log $TMP_PATH $TMP_BIN_PATH $TMP_SCRIPT_FUNC_PATH $TMP_ROUTE_PATH $TMP_ACL_PATH $TMP_PATH2
get_config
export V2RAY_LOCATION_ASSET=$(config_t_get global_rules v2ray_location_asset "/usr/share/v2ray/")
export XRAY_LOCATION_ASSET=$V2RAY_LOCATION_ASSET
export ENABLE_DEPRECATED_GEOSITE=true
export ENABLE_DEPRECATED_GEOIP=true
ulimit -n 65535
start_haproxy
start_socks
nftflag=0
local use_nft=$(config_t_get global_forwarding use_nft 0)
local USE_TABLES
if [ "$use_nft" == 0 ]; then
if [ -n "$(command -v iptables-legacy || command -v iptables)" ] && [ -n "$(command -v ipset)" ] && [ -n "$(dnsmasq --version | grep 'Compile time options:.* ipset')" ]; then
USE_TABLES="iptables"
else
echolog "系统未安装iptables或ipset或Dnsmasq没有开启ipset支持,无法使用iptables+ipset透明代理!"
if [ -n "$(command -v fw4)" ] && [ -n "$(command -v nft)" ] && [ -n "$(dnsmasq --version | grep 'Compile time options:.* nftset')" ]; then
echolog "检测到fw4,使用nftables进行透明代理。"
USE_TABLES="nftables"
nftflag=1
config_t_set global_forwarding use_nft 1
uci -q commit ${CONFIG}
fi
fi
else
if [ -n "$(dnsmasq --version | grep 'Compile time options:.* nftset')" ]; then
USE_TABLES="nftables"
nftflag=1
else
echolog "Dnsmasq软件包不满足nftables透明代理要求,如需使用请确保dnsmasq版本在2.87以上并开启nftset支持。"
fi
fi
check_depends $USE_TABLES
[ "$USE_TABLES" = "nftables" ] && {
dnsmasq_version=$(dnsmasq -v | grep -i "Dnsmasq version " | awk '{print $3}')
[ "$(expr $dnsmasq_version \>= 2.90)" == 0 ] && echolog "Dnsmasq版本低于2.90,建议升级至2.90及以上版本以避免部分情况下Dnsmasq崩溃问题!"
}
if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then
[ "$(uci -q get dhcp.@dnsmasq[0].dns_redirect)" == "1" ] && {
uci -q set ${CONFIG}.@global[0].dnsmasq_dns_redirect='1'
uci -q commit ${CONFIG}
uci -q set dhcp.@dnsmasq[0].dns_redirect='0'
uci -q commit dhcp
lua $APP_PATH/helper_dnsmasq.lua restart -LOG 0
}
fi
[ "$ENABLED_DEFAULT_ACL" == 1 ] && {
mkdir -p ${GLOBAL_ACL_PATH}
start_redir TCP
start_redir UDP
start_dns
}
[ -n "$USE_TABLES" ] && source $APP_PATH/${USE_TABLES}.sh start
set_cache_var "USE_TABLES" "$USE_TABLES"
if [ "$ENABLED_DEFAULT_ACL" == 1 ] || [ "$ENABLED_ACLS" == 1 ]; then
bridge_nf_ipt=$(sysctl -e -n net.bridge.bridge-nf-call-iptables)
set_cache_var "bak_bridge_nf_ipt" "$bridge_nf_ipt"
sysctl -w net.bridge.bridge-nf-call-iptables=0 >/dev/null 2>&1
[ "$PROXY_IPV6" == "1" ] && {
bridge_nf_ip6t=$(sysctl -e -n net.bridge.bridge-nf-call-ip6tables)
set_cache_var "bak_bridge_nf_ip6t" "$bridge_nf_ip6t"
sysctl -w net.bridge.bridge-nf-call-ip6tables=0 >/dev/null 2>&1
}
fi
start_crontab
echolog "运行完成!\n"
}
stop() {
clean_log
eval_cache_var
[ -n "$USE_TABLES" ] && source $APP_PATH/${USE_TABLES}.sh stop
delete_ip2route
# 结束 SS 插件进程
# kill_all xray-plugin v2ray-plugin obfs-local shadow-tls
local pid_file pid
find "$TMP_PATH" -type f -name '*_plugin.pid' | while read -r pid_file; do
read -r pid < "$pid_file"
if [ -n "$pid" ]; then
kill -9 "$pid" >/dev/null 2>&1
fi
done
pgrep -f "sleep.*(6s|9s|58s)" | xargs kill -9 >/dev/null 2>&1
pgrep -af "${CONFIG}/" | awk '! /app\.sh|subscribe\.lua|rule_update\.lua|tasks\.sh|ujail/{print $1}' | xargs kill -9 >/dev/null 2>&1
stop_crontab
source $APP_PATH/helper_smartdns.sh del
rm -rf $GLOBAL_DNSMASQ_CONF
rm -rf $GLOBAL_DNSMASQ_CONF_PATH
[ "1" = "1" ] && {
#restore logic
bak_dnsmasq_dns_redirect=$(config_t_get global dnsmasq_dns_redirect)
[ -n "${bak_dnsmasq_dns_redirect}" ] && {
uci -q set dhcp.@dnsmasq[0].dns_redirect="${bak_dnsmasq_dns_redirect}"
uci -q commit dhcp
uci -q delete ${CONFIG}.@global[0].dnsmasq_dns_redirect
uci -q commit ${CONFIG}
}
if [ -z "${ACL_default_dns_port}" ] || [ -n "${bak_dnsmasq_dns_redirect}" ]; then
uci -q del_list dhcp.@dnsmasq[0].addnmount="${GLOBAL_DNSMASQ_CONF_PATH}"
uci -q commit dhcp
lua $APP_PATH/helper_dnsmasq.lua restart -LOG 0
fi
[ -n "${bak_bridge_nf_ipt}" ] && sysctl -w net.bridge.bridge-nf-call-iptables=${bak_bridge_nf_ipt} >/dev/null 2>&1
[ -n "${bak_bridge_nf_ip6t}" ] && sysctl -w net.bridge.bridge-nf-call-ip6tables=${bak_bridge_nf_ip6t} >/dev/null 2>&1
}
rm -rf $TMP_PATH
rm -rf /tmp/lock/${CONFIG}_socks_auto_switch*
rm -rf /tmp/lock/${CONFIG}_lease2hosts*
echolog "清空并关闭相关程序和缓存完成。"
exit 0
}
get_config() {
ENABLED_DEFAULT_ACL=0
TCP_REDIR_PORT=1041
UDP_REDIR_PORT=1051
ENABLED=$(config_t_get global enabled 0)
SOCKS_ENABLED=$(config_t_get global socks_enabled 0)
TCP_NODE=$(config_t_get global tcp_node)
UDP_NODE=$(config_t_get global udp_node)
TCP_UDP=0
if [ "$UDP_NODE" == "tcp" ]; then
UDP_NODE=$TCP_NODE
TCP_UDP=1
elif [ "$UDP_NODE" == "$TCP_NODE" ]; then
TCP_UDP=1
fi
[ "$ENABLED" == 1 ] && {
[ -n "$TCP_NODE" ] && [ "$(config_get_type $TCP_NODE)" == "nodes" ] && ENABLED_DEFAULT_ACL=1
[ -n "$UDP_NODE" ] && [ "$(config_get_type $UDP_NODE)" == "nodes" ] && ENABLED_DEFAULT_ACL=1
}
ENABLED_ACLS=$(config_t_get global acl_enable 0)
[ "$ENABLED_ACLS" == 1 ] && {
[ "$(uci show ${CONFIG} | grep "@acl_rule" | grep "enabled='1'" | wc -l)" == 0 ] && ENABLED_ACLS=0
}
TCP_PROXY_WAY=$(config_t_get global_forwarding tcp_proxy_way redirect)
PROXY_IPV6=$(config_t_get global_forwarding ipv6_tproxy 0)
TCP_REDIR_PORTS=$(config_t_get global_forwarding tcp_redir_ports '80,443')
UDP_REDIR_PORTS=$(config_t_get global_forwarding udp_redir_ports '1:65535')
TCP_NO_REDIR_PORTS=$(config_t_get global_forwarding tcp_no_redir_ports 'disable')
UDP_NO_REDIR_PORTS=$(config_t_get global_forwarding udp_no_redir_ports 'disable')
TCP_PROXY_DROP_PORTS=$(config_t_get global_forwarding tcp_proxy_drop_ports 'disable')
UDP_PROXY_DROP_PORTS=$(config_t_get global_forwarding udp_proxy_drop_ports '80,443')
USE_DIRECT_LIST=$(config_t_get global use_direct_list 1)
USE_PROXY_LIST=$(config_t_get global use_proxy_list 1)
USE_BLOCK_LIST=$(config_t_get global use_block_list 1)
USE_GFW_LIST=$(config_t_get global use_gfw_list 1)
CHN_LIST=$(config_t_get global chn_list direct)
TCP_PROXY_MODE=$(config_t_get global tcp_proxy_mode proxy)
UDP_PROXY_MODE=$(config_t_get global udp_proxy_mode proxy)
[ "${TCP_PROXY_MODE}" != "disable" ] && TCP_PROXY_MODE="proxy"
[ "${UDP_PROXY_MODE}" != "disable" ] && UDP_PROXY_MODE="proxy"
LOCALHOST_PROXY=$(config_t_get global localhost_proxy 1)
[ "${LOCALHOST_PROXY}" == 1 ] && {
LOCALHOST_TCP_PROXY_MODE=$TCP_PROXY_MODE
LOCALHOST_UDP_PROXY_MODE=$UDP_PROXY_MODE
}
CLIENT_PROXY=$(config_t_get global client_proxy 1)
DNS_SHUNT=$(config_t_get global dns_shunt dnsmasq)
[ -z "$(first_type $DNS_SHUNT)" ] && DNS_SHUNT="dnsmasq"
DNS_MODE=$(config_t_get global dns_mode tcp)
[ "$DNS_SHUNT" = "smartdns" ] && DNS_MODE=$(config_t_get global smartdns_dns_mode socks)
REMOTE_DNS=$(config_t_get global remote_dns 1.1.1.1:53 | sed 's/#/:/g' | sed -E 's/\:([^:]+)$/#\1/g')
USE_DEFAULT_DNS=$(config_t_get global use_default_dns direct)
FILTER_PROXY_IPV6=$(config_t_get global filter_proxy_ipv6 0)
DNS_REDIRECT=$(config_t_get global dns_redirect 1)
REDIRECT_LIST="socks ss ss-rust ssr sing-box xray trojan-plus naiveproxy hysteria2"
TPROXY_LIST="socks ss ss-rust ssr sing-box xray trojan-plus hysteria2"
NEXT_DNS_LISTEN_PORT=15353
TUN_DNS="127.0.0.1#${NEXT_DNS_LISTEN_PORT}"
DNS_CACHE=0
DIRECT_DNS_QUERY_STRATEGY="UseIP"
REMOTE_DNS_QUERY_STRATEGY="UseIP"
[ "$FILTER_PROXY_IPV6" = "1" ] && REMOTE_DNS_QUERY_STRATEGY="UseIPv4"
DNSMASQ_FILTER_PROXY_IPV6=${FILTER_PROXY_IPV6}
RESOLVFILE=/tmp/resolv.conf.d/resolv.conf.auto
[ -f "${RESOLVFILE}" ] && [ -s "${RESOLVFILE}" ] || RESOLVFILE=/tmp/resolv.conf.auto
ISP_DNS=$(cat $RESOLVFILE 2>/dev/null | grep -E -o "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+" | grep -v -E '^(0\.0\.0\.0|127\.0\.0\.1)$' | awk '!seen[$0]++')
ISP_DNS6=$(cat $RESOLVFILE 2>/dev/null | grep -E "([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4}" | awk -F % '{print $1}' | awk -F " " '{print $2}' | grep -v -Fx ::1 | grep -v -Fx :: | awk '!seen[$0]++')
DEFAULT_DNS=$(uci show dhcp.@dnsmasq[0] | grep "\.server=" | awk -F '=' '{print $2}' | sed "s/'//g" | tr ' ' '\n' | grep -v "\/" | head -2 | sed ':label;N;s/\n/,/;b label')
[ -z "${DEFAULT_DNS}" ] && [ "$(echo $ISP_DNS | tr ' ' '\n' | wc -l)" -le 2 ] && DEFAULT_DNS=$(echo -n $ISP_DNS | tr ' ' '\n' | head -2 | tr '\n' ',' | sed 's/,$//')
LOCAL_DNS="${DEFAULT_DNS:-119.29.29.29,223.5.5.5}"
IPT_APPEND_DNS=${LOCAL_DNS}
DNSMASQ_CONF_DIR=/tmp/dnsmasq.d
DEFAULT_DNSMASQ_CFGID="$(uci -q show "dhcp.@dnsmasq[0]" | awk 'NR==1 {split($0, conf, /[.=]/); print conf[2]}')"
if [ -f "/tmp/etc/dnsmasq.conf.$DEFAULT_DNSMASQ_CFGID" ]; then
DNSMASQ_CONF_DIR="$(awk -F '=' '/^conf-dir=/ {print $2}' "/tmp/etc/dnsmasq.conf.$DEFAULT_DNSMASQ_CFGID")"
if [ -n "$DNSMASQ_CONF_DIR" ]; then
DNSMASQ_CONF_DIR=${DNSMASQ_CONF_DIR%*/}
else
DNSMASQ_CONF_DIR="/tmp/dnsmasq.d"
fi
fi
set_cache_var GLOBAL_DNSMASQ_CONF ${DNSMASQ_CONF_DIR}/dnsmasq-${CONFIG}.conf
set_cache_var GLOBAL_DNSMASQ_CONF_PATH ${GLOBAL_ACL_PATH}/dnsmasq.d
}
arg1=$1
shift
case $arg1 in
add_ip2route)
add_ip2route $@
;;
echolog)
echolog $@
;;
get_new_port)
get_new_port $@
;;
get_cache_var)
get_cache_var $@
;;
set_cache_var)
set_cache_var $@
;;
run_socks)
run_socks $@
;;
socks_node_switch)
socks_node_switch $@
;;
start)
start
;;
stop)
stop
;;
esac
|
294coder/Efficient-MIF | 22,774 | utils/train_utils.py | from pathlib import Path
import sys
sys.path.append("./")
import zipfile
from io import BytesIO
from pathlib import Path
from PIL import Image
import PIL.Image as Image
from contextlib import contextmanager
from collections import OrderedDict
from typing import Union, TYPE_CHECKING
import h5py
import numpy as np
import torch
from torch.utils.data import DataLoader
from utils import h5py_to_dict, NameSpace, easy_logger
if TYPE_CHECKING:
from accelerate import Accelerator
def get_eval_dataset(args, logger=None):
from task_datasets.WV3 import WV3Datasets
from task_datasets.GF2 import GF2Datasets
from task_datasets.HISR import HISRDatasets
from task_datasets.TNO import TNODataset
from task_datasets.RoadScene import RoadSceneDataset
from task_datasets.LLVIP import LLVIPDALIPipeLoader
from task_datasets.MSRS import MSRSDatasets
from task_datasets.M3FD import M3FDDALIPipeLoader
from task_datasets.MedHarvard import MedHarvardDataset
from task_datasets.SICE import SICEDataset
logger = easy_logger(func_name='get_eval_dataset')
val_ds, val_dl = None, None
logger.info(f"use dataset: {args.dataset} on VIS-IR fusion task")
# 1. vis-ir image fusion (without gt)
if args.dataset == "roadscene":
val_ds = RoadSceneDataset(
args.path.base_dir, "test", no_split=True, get_name=True
)
elif args.dataset == "tno":
val_ds = TNODataset(
args.path.base_dir, "test", aug_prob=0.0, no_split=True, get_name=True
)
elif args.dataset == "msrs":
val_ds = MSRSDatasets(
args.path.base_dir,
mode=args.dataset_mode, # or 'test'/'detection'
transform_ratio=0.0,
get_name=True,
reduce_label=args.reduce_label,
)
elif args.dataset == "llvip":
val_dl = LLVIPDALIPipeLoader(
args.path.base_dir,
"test",
batch_size=args.val_bs,
device=args.device,
shuffle=False,
with_mask=True,
get_name=True,
reduce_label=args.reduce_label,
)
elif args.dataset == "m3fd":
val_dl = M3FDDALIPipeLoader(
args.path.base_dir,
"test",
batch_size=args.val_bs,
device=args.device,
shuffle=False,
with_mask=True,
get_name=True,
reduce_label=args.reduce_label,
)
elif args.dataset == "med_harvard":
val_ds = MedHarvardDataset(
args.path.base_dir,
mode="test",
device=args.device,
data_source="xmu",
get_name=True,
task="SPECT-MRI",
)
elif args.dataset == "sice":
val_ds = SICEDataset(
data_dir=args.path.base_dir,
mode="test",
transformer_ratio=0.0,
only_y=args.only_y,
get_name=True,
)
## 2. sharpening datasets (with gt)
elif args.dataset in [
"wv3",
"qb",
"gf2",
"cave_x4",
"harvard_x4",
"cave_x8",
"harvard_x8",
"hisi-houston",
"pavia",
"chikusei",
"botswana",
]:
# the dataset has already splitted
logger.info(f"use dataset: {args.dataset} on pansharpening/HISR task")
# FIXME: 需要兼顾老代码(只有trian_path和val_path)的情况
if hasattr(args.path, "val_path"):
# 旧代码:手动切换数据集路径
val_path = args.path.val_path
else:
_args_path_keys = list(args.path.__dict__.keys())
for k in _args_path_keys:
if args.dataset in k:
val_path = getattr(args.path, f"{args.dataset}_val_path")
assert val_path is not None, "val_path should not be None"
if val_path is not None:
assert val_path.endswith(".h5"), 'val_path should end with ".h5"'
h5_val = h5py.File(val_path)
# 1. parsharpening
if args.dataset in ["wv3", "qb"]:
d_val = h5py_to_dict(h5_val)
val_ds = WV3Datasets(d_val, hp=args.hp, aug_prob=0.0)
elif args.dataset == "gf2":
d_val = h5py_to_dict(h5_val)
val_ds = GF2Datasets(d_val, hp=args.hp, aug_prob=0.0)
# 2. hyperspectral image fusion
elif (
args.dataset[:4] == "cave"
or args.dataset[:7] == "harvard"
or args.dataset[:8] == "chikusei"
or args.dataset[:5] == "pavia"
or args.dataset[:8] == "botswana"
):
keys = ["LRHSI", "HSI_up", "RGB", "GT"]
if args.dataset.split("-")[-1] == "houston":
from einops import rearrange
# to avoid unpicklable error
def permute_fn(x):
return rearrange(x, "b h w c -> b c h w")
dataset_fn = permute_fn
else:
dataset_fn = None
d_val = h5py_to_dict(h5_val, keys)
val_ds = HISRDatasets(
d_val, aug_prob=args.aug_probs[1], dataset_fn=dataset_fn
)
# del h5_train, h5_val
else:
raise NotImplementedError(f"not support dataset {args.dataset}")
return val_ds, val_dl
def get_fusion_dataset(
args: NameSpace, accelerator: "Accelerator", device: Union[str, torch.device]
):
logger = easy_logger()
train_ds, val_ds, train_dl, val_dl = None, None, None, None
if args.dataset in [
"flir",
"tno",
"roadscene_tno_joint",
"vis_ir_joint",
"msrs",
"llvip",
"med_harvard",
"m3fd",
"sice",
"mefb",
]:
args.task = "fusion"
args.has_gt = False
args.path.base_dir = getattr(args.path, f"{args.dataset}_base_dir")
if args.dataset == "roadscene":
from task_datasets.RoadScene import RoadSceneDataset
train_ds = RoadSceneDataset(args.path.base_dir, "train")
val_ds = RoadSceneDataset(args.path.base_dir, "test")
elif args.dataset in ["tno", "roadscene_tno_joint"]:
from task_datasets.TNO import TNODataset
train_ds = TNODataset(
args.path.base_dir,
"train",
aug_prob=args.aug_probs[0],
duplicate_vis_channel=True,
)
val_ds = TNODataset(
args.path.base_dir,
"test",
aug_prob=args.aug_probs[1],
no_split=True,
duplicate_vis_channel=True,
)
elif args.dataset == "msrs":
from task_datasets.MSRS import MSRSDatasets
train_ds = MSRSDatasets(
args.path.base_dir,
"train",
transform_ratio=args.aug_probs[0],
output_size=args.fusion_crop_size,
n_proc_load=1,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False, # args.only_y
)
val_ds = MSRSDatasets(
args.path.base_dir,
"test",
transform_ratio=args.aug_probs[1],
output_size=args.fusion_crop_size,
n_proc_load=1,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False, # args.only_y
)
elif args.dataset == "llvip":
from task_datasets.LLVIP import LLVIPDALIPipeLoader
# We use DALI pipeline to accelerate the data loading process
train_dl = LLVIPDALIPipeLoader(
args.path.base_dir,
"train",
batch_size=args.train_bs,
output_size=args.fusion_crop_size,
device=accelerator.device,
num_shards=accelerator.num_processes,
shard_id=accelerator.process_index,
shuffle=True,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False, # args.only_y
)
val_dl = LLVIPDALIPipeLoader(
args.path.base_dir,
"test",
batch_size=args.val_bs,
device=accelerator.device,
fast_eval_n_samples=args.fast_eval_n_samples,
num_shards=accelerator.num_processes,
shard_id=accelerator.process_index,
shuffle=True,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False, # args.only_y
)
elif args.dataset == "m3fd":
from task_datasets import M3FDDALIPipeLoader
train_dl = M3FDDALIPipeLoader(
args.path.base_dir,
"train",
batch_size=args.train_bs,
output_size=args.fusion_crop_size,
device=accelerator.device,
num_shards=accelerator.num_processes,
shard_id=accelerator.process_index,
shuffle=True,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False, # args.only_y
)
val_dl = M3FDDALIPipeLoader(
args.path.base_dir,
"test",
batch_size=args.val_bs,
device=accelerator.device,
fast_eval_n_samples=args.fast_eval_n_samples,
num_shards=accelerator.num_processes,
shard_id=accelerator.process_index,
shuffle=True,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False, # args.only_y
)
elif args.dataset == "vis_ir_joint":
from task_datasets import VISIRJointGenericLoader
train_dl = VISIRJointGenericLoader(
vars(args.path.base_dir),
mode="train",
batch_size=args.train_bs,
output_size=args.fusion_crop_size,
device=accelerator.device,
shuffle_in_dataset=True,
num_shards=accelerator.num_processes,
shard_id=accelerator.process_index,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False,
)
val_dl = VISIRJointGenericLoader(
## only test msrs and roadscene_tno_joint dataset
{'msrs': args.path.base_dir['msrs'],
'roadscene_tno_joint': args.path.base_dir['roadscene_tno_joint']},
mode="test",
output_size=224, # enforce the different size images to be the same size
batch_size=args.val_bs,
device=accelerator.device,
shuffle_in_dataset=False,
fast_eval_n_samples=30,
num_shards=accelerator.num_processes,
shard_id=accelerator.process_index,
reduce_label=args.datasets_cfg.reduce_label,
only_y_component=False,
)
elif args.dataset == "med_harvard":
from task_datasets.MedHarvard import MedHarvardDataset
if getattr(args, "datasets_cfg", None):
task = args.datasets_cfg.med_harvard.task
else:
task = None
train_ds = MedHarvardDataset(
args.path.base_dir,
mode="train",
device=device,
data_source="xmu",
transform_ratio=args.aug_probs[0],
task=task,
)
val_ds = MedHarvardDataset(
args.path.base_dir,
mode="test",
device=device,
data_source="xmu",
task=task,
)
assert args.num_worker == 0, "num_worker should be 0 for MedHarvard dataset"
elif args.dataset == "sice":
from task_datasets.SICE import SICEDataset
train_ds = SICEDataset(
data_dir=args.path.base_dir,
mode="train",
transformer_ratio=args.aug_probs[0],
output_size=args.fusion_crop_size,
only_y=args.only_y,
)
val_ds = SICEDataset(
data_dir=args.path.base_dir,
mode="train",
transformer_ratio=args.aug_probs[0],
output_size=args.fusion_crop_size,
only_y=args.only_y,
)
args.has_gt = True
else:
raise NotImplementedError(f"not support dataset {args.dataset}")
elif args.dataset in [
"wv3",
"qb",
"gf2",
"cave_x4",
"harvard_x4",
"cave_x8",
"harvard_x8",
"hisi-houston",
]:
args.task = "sharpening"
# the dataset has already splitted
# FIXME: 需要兼顾老代码(只有trian_path和val_path)的情况
if hasattr(args.path, "train_path") and hasattr(args.path, "val_path"):
# 旧代码:手动切换数据集路径
train_path = args.path.train_path
val_path = args.path.val_path
else:
_args_path_keys = list(args.path.__dict__.keys())
for k in _args_path_keys:
if args.dataset in k:
train_path = getattr(args.path, f"{args.dataset}_train_path")
val_path = getattr(args.path, f"{args.dataset}_val_path")
assert (
train_path is not None and val_path is not None
), "train_path and val_path should not be None"
h5_train, h5_val = (
h5py.File(train_path),
h5py.File(val_path),
)
if args.dataset in ["wv3", "qb"]:
from task_datasets.WV3 import WV3Datasets, make_datasets
d_train, d_val = h5py_to_dict(h5_train), h5py_to_dict(h5_val)
train_ds, val_ds = (
WV3Datasets(d_train, aug_prob=args.aug_probs[0]),
WV3Datasets(d_val, aug_prob=args.aug_probs[1]),
)
elif args.dataset == "gf2":
from task_datasets.GF2 import GF2Datasets
d_train, d_val = h5py_to_dict(h5_train), h5py_to_dict(h5_val)
train_ds, val_ds = (
GF2Datasets(d_train, aug_prob=args.aug_probs[0]),
GF2Datasets(d_val, aug_prob=args.aug_probs[1]),
)
elif args.dataset[:4] == "cave" or args.dataset[:7] == "harvard":
from task_datasets.HISR import HISRDatasets
keys = ["LRHSI", "HSI_up", "RGB", "GT"]
if args.dataset.split("-")[-1] == "houston":
from einops import rearrange
def permute_fn(x):
return rearrange(x, "b h w c -> b c h w")
dataset_fn = permute_fn
else:
dataset_fn = None
d_train, d_val = (
h5py_to_dict(h5_train, keys),
h5py_to_dict(h5_val, keys),
)
train_ds = HISRDatasets(
d_train, aug_prob=args.aug_probs[0], dataset_fn=dataset_fn
)
val_ds = HISRDatasets(
d_val, aug_prob=args.aug_probs[1], dataset_fn=dataset_fn
)
# del h5_train, h5_val
else:
raise NotImplementedError(f"not support dataset {args.dataset}")
train_sampler, val_sampler = None, None
if train_dl is None:
train_dl = DataLoader(
train_ds,
args.train_bs,
num_workers=args.num_worker,
sampler=train_sampler,
prefetch_factor=8 if args.num_worker > 0 else None,
pin_memory=False,
shuffle=args.shuffle if not args.ddp else None,
drop_last=True if args.shuffle else False,
)
if val_dl is None:
val_dl = DataLoader(
val_ds,
args.val_bs, # assert bs is 1, when using PatchMergeModule
num_workers=0,
sampler=val_sampler,
pin_memory=False,
shuffle=args.shuffle if not args.ddp else None,
drop_last=False,
)
return train_ds, train_dl, val_ds, val_dl
def set_ema_model_params_with_keys(ema_model_params: "dict[str, list[torch.Tensor] | int | float]",
keys: "list[str]",
keys_set: list[str]=['shadow_params']):
"""set ema model parameters with keys
Args:
ema_model_params (dict[str, list[torch.Tensor] | int | float]): ema model parameters
keys (list[str]): keys
Returns:
dict: ema model parameters with keys
"""
logger = easy_logger()
if not isinstance(keys, list):
keys = list(keys)
ema_model_params_with_keys = OrderedDict()
for k in ema_model_params.keys():
if k in keys_set and k in ema_model_params:
logger.info(f'set ema_model {k} params with keys')
params = ema_model_params[k]
assert params is not None
assert len(params) == len(keys), "ema_model_params and keys should have the same length"
_params = OrderedDict()
for mk, p in zip(keys, params):
_params[mk] = p
ema_model_params_with_keys[k] = _params
elif k not in keys_set and k in ema_model_params:
ema_model_params_with_keys[k] = ema_model_params[k]
return ema_model_params_with_keys
def run_once(abled=True):
def _inner(func):
def _wrapper(*args, **kwargs):
nonlocal abled
if not abled:
return None
else:
outs = func(*args, **kwargs)
abled = False
return outs
return _wrapper
return _inner
def sanity_check(func: callable):
@run_once()
def _inner(*args, **kwargs):
return func(*args, **kwargs)
return _inner
@contextmanager
def save_imgs_in_zip(
zipfile_name: str, mode="w", verbose: bool = False, save_file_ext: str = "jpeg"
):
"""save images to a zip file
Args:
zipfile_name (str): zip filename
mode (str, optional): mode to write in. Defaults to "w".
verbose (bool, optional): print out. Defaults to False.
save_file_ext (str, optional): image extension in the zip file. Defaults to "jpeg".
Yields:
callable: a function to save image
Examples::
with save_imgs_in_zip('zip_file.zip') as add_image:
img, img_name = get_img()
add_image(img, img_name)
:ref: `add_image`
"""
logger = easy_logger()
# save_file_ext = save_file_ext.upper()
zf = zipfile.ZipFile(
zipfile_name, mode=mode, compression=zipfile.ZIP_DEFLATED, compresslevel=9
)
bytes_io = BytesIO()
# jpg compression
_jpg_quality = 100 # 95 if save_file_ext in ["jpeg", "jpg", "JPG", "JPEG"] else 100
try:
logger.info(f"zip file will be saved at {zipfile_name}")
def to_bytes(image_data, image_name):
batched_image_bytes = []
if image_data.ndim == 4: # batched rgb images
assert isinstance(image_name, list), "image_name should be a list"
assert image_data.shape[0] == len(
image_name
), "image_name should have the same length as image_data"
for img in image_data: # [b, h, w, c]
Image.fromarray(img).save(
bytes_io, format=save_file_ext, quality=_jpg_quality
)
batched_image_bytes.append(bytes_io.getvalue())
elif image_data.ndim == 3:
if image_data.shape[-1] == 1: # gray image # [h, w, 1]
Image.fromarray(image_data[..., 0]).save(
bytes_io, format=save_file_ext, quality=_jpg_quality
)
image_data = bytes_io.getvalue()
elif image_data.shape[-1] == 3:
Image.fromarray(image_data).save(
bytes_io, format=save_file_ext, quality=_jpg_quality
)
image_data = bytes_io.getvalue()
else:
raise ValueError(
f"image_data shape {image_data.shape} not supported"
)
elif image_data.ndim == 2: # gray image # [h, w]
Image.fromarray(image_data).save(
bytes_io, format=save_file_ext, quality=_jpg_quality
)
image_data = bytes_io.getvalue()
return image_data, batched_image_bytes
def add_image(
image_data: "Image.Image | np.ndarray | torch.Tensor | bytes",
image_name: "Union[str, list[str]]",
):
"""add image to the zipfile
Args:
image_data (Image.Image | np.ndarray | torch.Tensor | bytes): can be Image.Image, np.ndarray, torch.Tensor, bytes,
shape should be [b, h, w, c], [h, w, c], [h, w, 1]
image_name (str | list[str]): saved image names
"""
# to bytes
batched_image_bytes = None
if isinstance(image_data, Image.Image):
image_data.save(bytes_io, format=save_file_ext)
bytes = bytes_io.getvalue()
elif isinstance(image_data, np.ndarray):
bytes, batched_image_bytes = to_bytes(image_data, image_name)
elif isinstance(image_data, torch.Tensor):
image_data = image_data.detach().cpu().numpy()
bytes, batched_image_bytes = to_bytes(image_data, image_name)
else:
raise ValueError(f"image_data type {type(image_data)} not supported")
# saving to zip file
if batched_image_bytes is not None:
for i, img_bytes in enumerate(batched_image_bytes):
zf.writestr(image_name[i], img_bytes)
else:
zf.writestr(image_name, bytes)
if verbose:
logger.info(f"add image {image_name} to zip file")
bytes_io.seek(0)
bytes_io.truncate()
yield add_image
except Exception as e:
if verbose:
logger.error(e, raise_error=True)
raise e
finally:
if verbose:
logger.info(f"zip file saved at {zipfile_name}, zipfile close")
zf.close()
bytes_io.close()
|
2977094657/DidaAPI | 2,118 | frontend/docs/api/pomodoros/focus-distribution.md | # 获取专注详情分布
获取指定日期范围内的专注时长分布统计,包括按项目、标签、任务分布的专注时长。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/pomodoros/statistics/dist/{start_date}/{end_date}`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单官方API
## 请求参数
### 路径参数
| 参数名 | 类型 | 必填 | 描述 | 示例 |
|--------|------|------|------|------|
| start_date | string | 是 | 开始日期,格式: YYYYMMDD | 20231201 |
| end_date | string | 是 | 结束日期,格式: YYYYMMDD | 20231207 |
## 完整请求示例
```http
GET https://api.dida365.com/api/v2/pomodoros/statistics/dist/20231201/20231207 HTTP/1.1
Host: api.dida365.com
Cookie: t=string; _csrf_token=string
X-CSRFToken: string
```
## 响应格式
### 成功响应
**状态码**: `200 OK`
**响应体**:
```json
{
"projectDurations": {
"project_id_1": {
"duration": 3600,
"name": "工作项目",
"color": "#FF5722"
},
"project_id_2": {
"duration": 1800,
"name": "学习项目",
"color": "#2196F3"
}
},
"tagDurations": {
"tag_id_1": {
"duration": 2400,
"name": "编程",
"color": "#4CAF50"
},
"tag_id_2": {
"duration": 1200,
"name": "阅读",
"color": "#FF9800"
}
},
"taskDurations": {
"task_id_1": {
"duration": 1800,
"title": "完成API文档",
"projectId": "project_id_1"
},
"task_id_2": {
"duration": 1200,
"title": "学习新技术",
"projectId": "project_id_2"
}
}
}
```
## 响应字段说明
### projectDurations (按项目分布)
- **类型**: Object
- **描述**: 按项目分组的专注时长统计
- **字段说明**:
- `duration`: 专注时长(秒)
- `name`: 项目名称
- `color`: 项目颜色
### tagDurations (按标签分布)
- **类型**: Object
- **描述**: 按标签分组的专注时长统计
- **字段说明**:
- `duration`: 专注时长(秒)
- `name`: 标签名称
- `color`: 标签颜色
### taskDurations (按任务分布)
- **类型**: Object
- **描述**: 按任务分组的专注时长统计
- **字段说明**:
- `duration`: 专注时长(秒)
- `title`: 任务标题
- `projectId`: 所属项目ID
## 使用说明
1. **日期格式**: 必须使用 YYYYMMDD 格式
2. **时长单位**: 所有时长都以秒为单位
3. **数据范围**: 返回指定日期范围内的专注分布统计
4. **认证要求**: 需要有效的登录会话
## 相关接口
- [获取专注记录时间线](./focus-timeline.md) - 获取详细的专注记录
- [获取专注趋势热力图](./focus-heatmap.md) - 获取专注趋势数据
- [获取专注时间分布](./focus-time-distribution.md) - 获取时间段分布
- [获取专注时间按小时分布](./focus-hour-distribution.md) - 获取小时分布
|
2977094657/DidaAPI | 910 | frontend/docs/api/pomodoros/focus-time-distribution.md | # 获取专注时间分布
获取指定日期范围内按时间段分布的专注数据。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/pomodoros/statistics/dist/clockByDay/{start_date}/{end_date}`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
| 参数名 | 类型 | 必填 | 说明 | 示例 |
|--------|------|------|------|------|
| start_date | string | 是 | 开始日期,格式: YYYYMMDD | 20250526 |
| end_date | string | 是 | 结束日期,格式: YYYYMMDD | 20250601 |
## 响应格式
### 成功响应
```json
[
{
"day": "20250526",
"timezone": "Asia/Shanghai"
},
{
"timeDurations": {
"15": 60,
"22": 60,
"23": 60
},
"day": "20250601",
"timezone": "Asia/Shanghai"
}
]
```
### 响应字段说明
| 字段 | 类型 | 说明 |
|------|------|------|
| day | string | 日期(YYYYMMDD格式) |
| timezone | string | 时区 |
| timeDurations | object | 时间段专注分布(可选字段) |
| timeDurations.{hour} | number | 指定小时的专注时长(分钟),hour为24小时制 |
|
294coder/Efficient-MIF | 11,517 | utils/inference_helper_func.py | # GPL License
# Copyright (C) 2024 , UESTC
# All Rights Reserved
#
# @Time : 2021/10/15 17:53
# @Author : Zihan Cao, Xiao Wu
from functools import partial
import inspect
from typing import Tuple, Optional
import einops
import torch
import torch.nn as nn
import torch.nn.functional as F
import math
from torch.utils.data import DataLoader
from torch import Tensor, nn
from tqdm import tqdm
import json
import re
import glob
from .visualize import viz_batch, res_image
from .metric_sharpening import AnalysisPanAcc
from typing import TYPE_CHECKING
if TYPE_CHECKING:
from model.base_model import BaseModel
def has_patch_merge_model(model: "nn.Module | BaseModel"):
return (hasattr(model, '_patch_merge_model')) or (hasattr(model, 'patch_merge_model'))
def patch_merge_in_val_step(model: "nn.Module | BaseModel"):
return 'patch_merge' in list(inspect.signature(model.val_step).parameters.keys())
def has_patch_merge_model(model: "nn.Module | BaseModel"):
return (hasattr(model, '_patch_merge_model')) or (hasattr(model, 'patch_merge_model'))
def patch_merge_in_val_step(model):
return 'patch_merge' in list(inspect.signature(model.val_step).parameters.keys())
# callback function
def basic_callback(model: "BaseModel", iter_idx: int):
from utils import get_local
assert get_local().cache is not None and get_local.is_activate
cache = get_local().cache
attns = cache['_forward_implem']
get_local.clear()
############ Main inference function ############
@torch.no_grad()
@torch.inference_mode()
def unref_for_loop(model,
dl: DataLoader,
device,
*,
split_patch=False,
feature_callback: callable=None,
**patch_merge_module_kwargs):
from model.base_model import PatchMergeModule
all_sr = []
try:
spa_size = tuple(dl.dataset.lms.shape[-2:])
except AttributeError:
spa_size = tuple(dl.dataset.rgb.shape[-2:])
inference_bar = tqdm(enumerate(dl, 1), dynamic_ncols=True, total=len(dl))
analysis = AnalysisPanAcc(ratio=patch_merge_module_kwargs.get('ergas_ratio', 4), ref=False,
sensor=patch_merge_module_kwargs.get('sensor', 'DEFAULT'),
default_max_value=patch_merge_module_kwargs.get('default_max_value', None))
if split_patch:
# check if has the patch merge model
if not (has_patch_merge_model(model) or patch_merge_in_val_step(model)):
# assert bs == 1, 'batch size should be 1'
# warp the model into PatchMergeModule
model = PatchMergeModule(net=model, device=device, **patch_merge_module_kwargs)
for i, (pan, ms, lms) in inference_bar:
pan, ms, lms = pan.to(device).float(), ms.to(device).float(), lms.to(device).float()
# split the image into several patches to avoid gpu OOM
if split_patch:
input = (ms, lms, pan)
if hasattr(model, 'forward_chop'):
sr = model.forward_chop(*input)[0]
elif patch_merge_in_val_step(model):
sr = model.val_step(*input, patch_merge=True)
else:
raise NotImplemented('model should have @forward_chop or patch_merge arg in @val_step')
else:
if patch_merge_in_val_step(model):
sr = model.val_step(ms, lms, pan, False)
else:
sr = model.val_step(ms, lms, pan)
sr = sr.clip(0, 1)
sr1 = sr.detach().cpu().numpy()
all_sr.append(sr1)
# analysis(sr, ms, lms, pan)
viz_batch(sr.detach().cpu(), suffix='sr', start_index=i, base_path='visualized_img/img_shows')
viz_batch(ms.detach().cpu(), suffix='ms', start_index=i, base_path='visualized_img/img_shows')
viz_batch(pan.detach().cpu(), suffix='pan', start_index=i, base_path='visualized_img/img_shows')
if feature_callback is not None:
feature_callback(model, i)
print(analysis.print_str())
return all_sr
@torch.no_grad()
@torch.inference_mode()
def ref_for_loop(model,
dl,
device,
*,
split_patch=False,
ergas_ratio=4,
residual_exaggerate_ratio=100,
feature_callback: callable=None,
**patch_merge_module_kwargs):
from model.base_model import PatchMergeModule
analysis = AnalysisPanAcc(ergas_ratio)
all_sr = []
inference_bar = tqdm(enumerate(dl, 1), dynamic_ncols=True, total=len(dl))
if not (has_patch_merge_model(model) or patch_merge_in_val_step(model)):
model = PatchMergeModule(net=model, device=device, **patch_merge_module_kwargs)
for i, (pan, ms, lms, gt) in inference_bar:
pan, ms, lms, gt = pan.to(device).float(), ms.to(device).float(), lms.to(device).float(), gt.to(device).float()
# split the image into several patches to avoid gpu OOM
if split_patch:
input = (ms, lms, pan)
if hasattr(model, 'forward_chop'):
sr = model.forward_chop(*input)[0]
elif patch_merge_in_val_step(model):
sr = model.val_step(*input, patch_merge=split_patch)
else:
raise NotImplemented('model should have @forward_chop or patch_merge arg in @val_step')
else:
if patch_merge_in_val_step(model):
sr = model.val_step(ms, lms, pan, False)
else:
sr = model.val_step(ms, lms, pan)
if feature_callback is not None:
feature_callback(model, i)
sr = sr.clip(0, 1)
sr1 = sr.detach().cpu().numpy()
all_sr.append(sr1)
analysis(gt, sr)
res = res_image(gt, sr, exaggerate_ratio=residual_exaggerate_ratio)
viz_batch(sr.detach().cpu(), suffix='sr', start_index=i, base_path='visualized_img/img_shows')
viz_batch(gt.detach().cpu(), suffix='gt', start_index=i, base_path='visualized_img/img_shows')
viz_batch(ms.detach().cpu(), suffix='ms', start_index=i, base_path='visualized_img/img_shows')
viz_batch(pan.detach().cpu(), suffix='pan', start_index=i, base_path='visualized_img/img_shows')
viz_batch(res.detach().cpu(), suffix='residual', start_index=i, base_path='visualized_img/img_shows')
print(analysis.print_str())
return all_sr
def find_data_path(dataset_type, full_res):
if dataset_type == "wv3":
if not full_res:
path = "/volsparse1/dataset/PanCollection/test_data/test_wv3_multiExm1.h5"
else:
path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_wv3_OrigScale_multiExm1.h5"
elif dataset_type == "cave":
path = "/Data2/ZiHanCao/datasets/HISI/new_cave/test_cave(with_up)x4.h5"
elif dataset_type == "cave_x8":
path = "/volsparse1/dataset/HISR/cave_x8/test_cave(with_up)x8_rgb.h5"
elif dataset_type == "harvard":
path = "/Data2/ShangqiDeng/data/HSI/harvard_x4/test_harvard(with_up)x4_rgb200.h5"
elif dataset_type == "harvard_x8":
path = "/volsparse1/dataset/HISR/harvard_x8/test_harvard(with_up)x8_rgb.h5"
elif dataset_type == "gf5":
if not full_res:
path = "/Data2/ZiHanCao/datasets/pansharpening/GF5-GF1/tap23/test_GF5_GF1_23tap_new.h5"
else:
path = "/Data2/ZiHanCao/datasets/pansharpening/GF5-GF1/tap23/test_GF5_GF1_OrigScale.h5"
elif dataset_type == "gf":
if not full_res:
path = "/Data2/ZiHanCao/datasets/pansharpening/gf/reduced_examples/test_gf2_multiExm1.h5"
else:
path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_gf2_OrigScale_multiExm1.h5"
elif dataset_type == "qb":
if not full_res:
path = "/Data2/ZiHanCao/datasets/pansharpening/qb/reduced_examples/test_qb_multiExm1.h5"
else:
path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_qb_OrigScale_multiExm1.h5"
elif dataset_type == "wv2":
if not full_res:
path = "/Data2/ZiHanCao/datasets/pansharpening/wv2/reduced_examples/test_wv2_multiExm1.h5"
else:
path = "/Data2/ZiHanCao/datasets/pansharpening/pansharpening_test/test_wv2_OrigScale_multiExm1.h5"
elif dataset_type == "roadscene":
path = "/Data2/ZiHanCao/datasets/RoadSceneFusion_1"
elif dataset_type == "tno":
path = "/Data2/ZiHanCao/datasets/TNO"
else:
raise NotImplementedError("not exists {} dataset".format(dataset_type))
return path
def find_key_args_in_log(arch, sub_arch, datasets, weight_path):
# handle weight_path
slash_with_id = re.findall(r'_[a-zA-Z0-9-]{8}(?=\.pth|_)', weight_path)[-1]
run_id = slash_with_id[1:]
if sub_arch is not None and sub_arch != '':
sub_arch = '_' + sub_arch
else:
sub_arch = ''
_log_path = f'log_file/{arch}{sub_arch}/{datasets}/*{run_id}*/config.json'
log_path = glob.glob(_log_path)
if len(log_path) != 1:
raise RuntimeError(f'>>> log file: {_log_path} not exists!')
print(f'>>> found run id: {log_path[0]} config')
args = json.loads(''.join(open(log_path[0], 'r').readlines()))
return args
def crop_inference(model: "BaseModel",
xs: Tuple[Tensor, Tensor, Tensor],
crop_size: Tuple[int] = (16, 64, 64),
stride: Tuple[int] = (8, 32, 32)):
# only support CAVE dataset
# input shape: 128, 512, 512
# xs: (hsi_lr, hsi_up, rgb)
torch.backends.cudnn.enable = True
torch.backends.cudnn.benchmark = True
# preprocessing
crop_xs = []
ncols = []
bs, out_c, _, _ = xs[0].shape
_, _, out_h, out_w = xs[-1].shape
for i in range(len(xs)):
x = xs[i]
_, c, h, _ = x.shape # assume h equals w
crop = crop_size[i]
s = stride[i]
ncol = (h - crop) // s
ncols.append(ncol)
crop_x = F.unfold(x, crop, stride=s)
crop_x = einops.rearrange(crop_x, 'b (c k l) m -> m b c k l', k=crop, l=crop, c=c)
crop_xs.append(crop_x)
# model inference
model.eval()
out = []
for i in range(crop_xs[0].size(0)):
input = [crop_xs[j][i].cuda(0) for j in range(len(xs))]
out.append(model.val_step(*input).detach().cpu()) # [bs * 225, 31, 64, 64]
del input
torch.cuda.empty_cache()
# input: 255*[b, 31, 64, 64]
out = torch.cat(out, dim=0)
# postprocessing
out = einops.rearrange(out, '(m b) c k l -> b (c k l) m', b=bs, k=crop_size[-1], l=crop_size[-1], c=out_c)
output = F.fold(out, output_size=(out_h, out_w),
kernel_size=(crop_size[-1], crop_size[-1]),
dilation=1,
padding=0,
stride=(stride[-1], stride[-1]))
return output
if __name__ == '__main__':
from model.dcformer_reduce import DCFormer_Reduce
model = DCFormer_Reduce(8, 'C').cuda(0)
ms = torch.randn(1, 8, 128, 128)
interp_ms = F.interpolate(ms, size=512)
lms = torch.randn(1, 8, 512, 512)
pan = torch.randn(1, 1, 512, 512)
expand_pan = pan.expand(-1, 8, -1, -1)
# print(model.val_step(ms, lms, pan).shape)
print(crop_inference(model, xs=(ms, lms, pan)).shape)
|
297854895/vue-tsx-admin | 2,780 | .eslintrc.js | module.exports = {
root: true,
env: {
node: true
},
extends: [
"plugin:vue/essential",
"@vue/typescript"
],
rules: {
"no-console": process.env.NODE_ENV === "production" ? "error" : "off",
"no-debugger": process.env.NODE_ENV === "production" ? "error" : "off",
"quotes": [2, "single"], //单引号
'new-cap': 'off',
'no-continue': 'off',
'no-empty': 'on',
'experimentalDecorators': 'off',
'consistent-return': 'off',
'no-plusplus': 'off',
'guard-for-in': 'off',
'prefer-destructuring': 'off',
'prefer-promise-reject-errors': 'off',
'camelcase': 'off',
'eslint-import-resolver-typescript': {
'extensions': ['.ts', '.tsx', '.d.ts']
},
'space-infix-ops': 'off',
'import/prefer-default-export': 'off',
'no-restricted-syntax': 0,
'no-restricted-globals': 'off',
'global-require': 'off',
'import/no-dynamic-require': 'off',
'import/no-unresolved': 'off',
'import/newline-after-import': 'off',
'import/extensions': 'off',
'no-param-reassign': 'off',
'class-methods-use-this': 'off',
'linebreak-style': 'off',
'no-console': 'off',
'no-debugger': 'off',
"no-console": 0, //不禁用console
// "no-debugger": 2, //禁用debugger
"no-var": 0, //对var警告
"semi": 0, //不强制使用分号
"no-irregular-whitespace": 0, //不规则的空白不允许
"no-trailing-spaces": 1, //一行结束后面有空格就发出警告
"eol-last": 0, //文件以单一的换行符结束
// "no-unused-vars": [2, {"vars": "local", "args": "after-used"}], //不能有声明后未被使用的变量或参数
"no-underscore-dangle": 0, //标识符不能以_开头或结尾
"no-alert": 2, //禁止使用alert confirm prompt
"no-lone-blocks": 0, //禁止不必要的嵌套块
"no-class-assign": 2, //禁止给类赋值
"no-cond-assign": 2, //禁止在条件表达式中使用赋值语句
"no-const-assign": 2, //禁止修改const声明的变量
"no-delete-var": 2, //不能对var声明的变量使用delete操作符
"no-dupe-keys": 2, //在创建对象字面量时不允许键重复
"no-duplicate-case": 2, //switch中的case标签不能重复
"no-dupe-args": 2, //函数参数不能重复
"no-empty": 2, //块语句中的内容不能为空
"no-func-assign": 2, //禁止重复的函数声明
"no-invalid-this": 0, //禁止无效的this,只能用在构造器,类,对象字面量
"no-redeclare": 2, //禁止重复声明变量
"no-spaced-func": 2, //函数调用时 函数名与()之间不能有空格
"no-this-before-super": 0, //在调用super()之前不能使用this或super
"no-undef": 2, //不能有未定义的变量
"no-use-before-define": 2, //未定义前不能使用
"jsx-quotes": [2, "prefer-double"], //强制在JSX属性(jsx-quotes)中一致使用双引号
'vue/attribute-hyphenation': [
'error',
'always'
],
'vue/html-end-tags': 'error',
'vue/html-indent': [
'error',
2
],
'vue/html-self-closing': 'error',
'vue/require-default-prop': 'error',
'vue/require-prop-types': 'error',
'vue/attributes-order': 'error',
'vue/order-in-components': 'error'
},
parserOptions: {
parser: "@typescript-eslint/parser"
}
};
|
281677160/openwrt-package | 27,599 | luci-app-passwall/root/usr/share/passwall/rules/chnroute6 | 2001:250::/30
2001:254::/31
2001:256:100::/48
2001:4510:400::/40
2001:678:10d0::/48
2001:678:120::/48
2001:67c:ebc::/48
2001:7fa:5::/48
2001:c68::/32
2001:cc0::/32
2001:da8::/32
2001:daa:1::/48
2001:daa:2::/47
2001:daa:4::/47
2001:daa:6::/48
2001:dc7::/32
2001:dd8:1::/48
2001:dd9::/48
2001:df0:ac40::/48
2001:df1:61c0::/48
2001:df3:3a80::/48
2001:df3:8b80::/48
2001:df7:1480::/48
2400:1160::/32
2400:3200::/32
2400:5280:f803::/48
2400:5400:10::/48
2400:5a60:100::/48
2400:5a60:2::/48
2400:5f60::/32
2400:6000::/32
2400:6460:300::/40
2400:6460:500::/40
2400:6460::/39
2400:6600::/32
2400:6e60:1301::/48
2400:7100::/32
2400:73e0::/32
2400:75aa::/32
2400:7bc0:20::/43
2400:7fc0:220::/44
2400:7fc0:240::/44
2400:7fc0:2a0::/44
2400:7fc0:2c0::/44
2400:7fc0:4000::/40
2400:7fc0:4100::/48
2400:7fc0:6000::/40
2400:7fc0:8000::/36
2400:7fc0::/40
2400:7fc0:a000::/36
2400:7fc0:bb00::/40
2400:7fc0:c000::/36
2400:8200::/32
2400:87c0::/32
2400:89c0:1010::/44
2400:89c0:1020::/44
2400:89c0:1050::/46
2400:89c0:1130::/44
2400:89c0:1150::/48
2400:89c0:2100::/48
2400:89c0:2200::/48
2400:89c0:3010::/44
2400:89c0:6000::/48
2400:89c0:6100::/48
2400:8fc0::/32
2400:9020:f010::/46
2400:9340::/32
2400:9380:8001::/48
2400:9380:8003::/48
2400:9380:8021::/48
2400:9380:8040::/48
2400:9380:8140::/48
2400:9380:8201::/48
2400:9380:8301::/48
2400:9380:9001::/48
2400:9380:9002::/48
2400:9380:9005::/48
2400:9380:9009::/48
2400:9380:900a::/48
2400:9380:9020::/47
2400:9380:9040::/47
2400:9380:9050::/47
2400:9380:9060::/48
2400:9380:9071::/48
2400:9380:9080::/47
2400:9380:90b0::/45
2400:9380:9100::/47
2400:9380:9121::/48
2400:9380:9201::/48
2400:9380:9202::/48
2400:9380:9220::/47
2400:9380:9240::/47
2400:9380:9250::/47
2400:9380:9260::/48
2400:9380:9271::/48
2400:9380:9272::/48
2400:9380:9280::/47
2400:9380:9282::/48
2400:9380:92b0::/45
2400:95e0::/48
2400:9600:8800::/48
2400:9620::/32
2400:a040::/32
2400:a860:1::/48
2400:a860:2::/47
2400:a860:4::/47
2400:a980::/29
2400:ae00:1981::/48
2400:b200::/32
2400:b600::/32
2400:b700::/48
2400:be00::/48
2400:cb80:e2e::/47
2400:cb80:e30::/44
2400:cb80:e40::/44
2400:da00::/32
2400:dd00::/28
2400:ebc0::/32
2400:ee00::/32
2400:f6e0::/32
2400:f720::/32
2400:fe00::/32
2401:1160::/32
2401:11a0:10::/44
2401:11a0:d150::/48
2401:11a0:d152::/48
2401:11a0:d158::/48
2401:1200::/48
2401:1320::/32
2401:1d40::/32
2401:20::/40
2401:2780::/32
2401:2e00::/32
2401:33c0::/32
2401:3480:2000::/48
2401:3480:3000::/36
2401:3480::/36
2401:34a0::/31
2401:3800::/32
2401:3880::/32
2401:3980::/32
2401:3a80::/32
2401:3b80::/32
2401:3c80::/32
2401:3d80::/32
2401:3e80::/32
2401:3f80::/32
2401:4180::/32
2401:4280::/32
2401:4380::/32
2401:4480::/32
2401:4580::/32
2401:4680::/32
2401:4780::/32
2401:4880::/32
2401:4a80::/32
2401:5c20:10::/48
2401:70e0::/32
2401:71c0::/48
2401:7660::/48
2401:7700::/32
2401:7d40::/32
2401:7e00::/32
2401:800::/32
2401:8be0::/48
2401:8d00:12::/48
2401:8d00:14::/48
2401:8d00:4::/48
2401:8d00:7::/48
2401:8d00:8::/47
2401:8d00::/46
2401:8d00:b::/48
2401:8d00:c::/48
2401:8d00:f::/48
2401:9a00:10::/46
2401:9a00::/44
2401:a140:1::/48
2401:b180::/32
2401:b400:11::/48
2401:b400:14::/48
2401:b400:16::/47
2401:b400:20::/47
2401:b400:8::/47
2401:b400::/45
2401:b680::/32
2401:bd60::/32
2401:be00::/32
2401:c020:14::/48
2401:c020:6::/48
2401:c020:8::/47
2401:ca00::/32
2401:cb80::/32
2401:cc00::/32
2401:ce00::/32
2401:d0e0:2::/47
2401:d180:10::/47
2401:d180:111::/48
2401:d180:2120::/48
2401:d180::/46
2401:d920::/48
2401:de00::/32
2401:ec00::/32
2401:f860:100::/40
2401:f860:86::/47
2401:f860:88::/47
2401:f860:90::/46
2401:f860:94::/47
2401:f860:a::/47
2401:f860:c::/46
2401:f860:f100::/40
2401:f860:f6::/48
2401:fa00:40::/43
2402:1440::/32
2402:2000::/32
2402:20e0:f000::/48
2402:3180:8000::/33
2402:3180::/46
2402:3c00::/32
2402:3f80:1400::/40
2402:4440::/32
2402:4b80::/32
2402:4e00::/32
2402:5e40::/32
2402:5ec0::/32
2402:6e80::/32
2402:6f40:2::/48
2402:6f40::/48
2402:6fc0::/48
2402:7040::/32
2402:7d80:240::/47
2402:7d80:6666::/48
2402:7d80:8888::/48
2402:7d80:9999::/48
2402:7d80::/48
2402:840:d000::/46
2402:840:e000::/46
2402:840:f000::/38
2402:8bc0::/32
2402:8cc0:200::/40
2402:8cc0::/40
2402:92c0::/48
2402:93c0:100::/48
2402:93c0:20::/48
2402:93c0::/48
2402:9a80::/32
2402:a200::/32
2402:b8c0:106::/48
2402:b8c0:186::/48
2402:b8c0:6::/48
2402:b8c0:86::/48
2402:b940:200::/39
2402:b940::/40
2402:c60::/48
2402:d340::/32
2402:db40:5100::/48
2402:db40:5900::/48
2402:db40:5f00::/46
2402:dfc0:50::/44
2402:dfc0::/44
2402:e380:100::/40
2402:e480::/32
2402:e740::/32
2402:e7c0::/32
2402:ef40::/32
2402:f000::/32
2402:f140:ff10::/46
2402:f140:ff14::/48
2402:f8c0:2::/48
2402:f8c0:5::/48
2402:f8c0::/47
2403:1b80::/48
2403:1ec0:1200::/48
2403:1ec0:1400::/48
2403:1ec0:1600::/48
2403:1ec0:1610::/48
2403:1ec0:1900::/48
2403:2040::/48
2403:2b40::/32
2403:3140::/32
2403:4240::/32
2403:4280::/47
2403:4300::/32
2403:4c80::/48
2403:5c80::/48
2403:600::/32
2403:6380:14::/47
2403:6380:40::/46
2403:6380:60::/44
2403:6a00::/32
2403:7580::/32
2403:8080:101::/48
2403:8c00::/32
2403:9b00::/32
2403:a100::/48
2403:a140:100::/40
2403:a140:10::/48
2403:a200::/32
2403:ac00::/32
2403:ad80:8008::/48
2403:b400::/32
2403:c80::/32
2403:c980::/32
2403:d400::/32
2403:f4c0::/48
2403:ffc0:1100::/40
2403:ffc0:1200::/39
2404:1c80::/32
2404:3700::/48
2404:4dc0::/32
2404:6380:1000::/48
2404:6380:8001::/48
2404:6380::/48
2404:6500:dcb3::/48
2404:7180:a000::/48
2404:7180:a010::/48
2404:7180:aa00::/48
2404:7180:aa10::/48
2404:7180:b001::/48
2404:7180:b002::/48
2404:7180:b010::/48
2404:7180:c001::/48
2404:7180:c002::/48
2404:7180:c011::/48
2404:7180:c012::/48
2404:7180:f000::/48
2404:7180:f010::/48
2404:7240::/33
2404:7600::/32
2404:7940::/32
2404:bc0:1::/48
2404:bc0:4000::/43
2404:bc0:4100::/43
2404:bc0:4200::/43
2404:bc0:4300::/44
2404:bc0:4400::/43
2404:bc0:4f00::/43
2404:c2c0:240::/44
2404:c2c0:280::/44
2404:c2c0:2c0::/44
2404:c2c0:4000::/40
2404:c2c0:4100::/48
2404:c2c0:501::/48
2404:c2c0:6000::/40
2404:c2c0:8000::/36
2404:c2c0::/40
2404:c2c0:bb00::/40
2404:c2c0:c000::/36
2404:c300::/32
2404:c940::/48
2404:e280::/47
2404:e5c0::/32
2404:e8c0::/32
2404:f4c0::/32
2405:1480:1000::/48
2405:1480:2000::/48
2405:1480:3000::/47
2405:3140:11::/48
2405:3140:21::/48
2405:3140:31::/48
2405:3140:3a::/48
2405:3bc0::/48
2405:66c0::/32
2405:68c0:21::/48
2405:6940::/48
2405:6c0:2::/48
2405:6c0:4::/48
2405:6f00:c101::/48
2405:6f00:c102::/48
2405:6f00:c170::/47
2405:6f00:c602::/48
2405:7040:6000::/47
2405:78c0:6e00::/43
2405:80:13::/48
2405:80:1::/48
2405:8280::/32
2405:8a40::/32
2405:a900:ffee::/48
2405:a900:fffe::/48
2405:ad00::/32
2405:b7c0::/32
2405:d900::/32
2405:e000::/32
2405:f580::/32
2405:f940::/32
2406:1080::/48
2406:1e40:f012::/47
2406:2700::/32
2406:280::/32
2406:3340::/32
2406:3640:1::/48
2406:3d80::/32
2406:4d00::/48
2406:52c0::/32
2406:5340:6666::/48
2406:5340:8888::/48
2406:5ac0::/32
2406:6100::/32
2406:840:8100::/40
2406:840:9000::/44
2406:840:9600::/43
2406:840:9620::/44
2406:840:9961::/48
2406:840:9962::/47
2406:840:996c::/48
2406:840:a7fe::/48
2406:840:a800::/37
2406:840:e080::/44
2406:840:e0cf::/48
2406:840:e0e0::/46
2406:840:e0e4::/47
2406:840:e0e8::/48
2406:840:e10f::/48
2406:840:e14f::/48
2406:840:e201::/48
2406:840:e230::/44
2406:840:e260::/48
2406:840:e2cf::/48
2406:840:e600::/45
2406:840:e608::/46
2406:840:e621::/48
2406:840:e666::/47
2406:840:e720::/44
2406:840:e770::/48
2406:840:e777::/48
2406:840:e80f::/48
2406:840:eb00::/46
2406:840:eb04::/47
2406:840:eb07::/48
2406:840:eb08::/48
2406:840:eb0b::/48
2406:840:eb0f::/48
2406:840:ee40::/47
2406:840:ee44::/48
2406:840:ee4b::/48
2406:840:ee4d::/48
2406:840:eee5::/48
2406:840:f200::/47
2406:840:f203::/48
2406:840:f380::/44
2406:840:f44f::/48
2406:840:fa01::/48
2406:840:fa40::/48
2406:840:fa60::/44
2406:840:fc10::/44
2406:840:fc20::/43
2406:840:fc40::/42
2406:840:fc80::/42
2406:840:fcc0::/44
2406:840:fcd0::/48
2406:840:fd40::/42
2406:840:fd80::/44
2406:840:fd9f::/48
2406:840:fda0::/43
2406:840:fdc0::/44
2406:840:fdd1::/48
2406:840:fde5::/48
2406:840:fde6::/47
2406:840:fe27::/48
2406:840:fe90::/46
2406:840:fe94::/48
2406:840:fe96::/47
2406:840:fe98::/46
2406:840:fe9d::/48
2406:840:fe9e::/48
2406:840:fea2::/47
2406:840:fea4::/46
2406:840:fea8::/45
2406:840:fec0::/48
2406:840:fec2::/47
2406:840:fec4::/47
2406:840:fec8::/48
2406:840:feca::/47
2406:840:fecc::/47
2406:840:fecf::/48
2406:840:fed1::/48
2406:840:fed8::/48
2406:840:fedb::/48
2406:840:fedc::/47
2406:840:fedf::/48
2406:840:fef0::/48
2406:840:fef3::/48
2406:840:fef4::/48
2406:840:fef6::/47
2406:840:fef8::/46
2406:840:fefc::/48
2406:840:feff::/48
2406:8880::/48
2406:b640:100::/48
2406:b640:4100::/48
2406:cac0:200::/40
2406:cac0::/40
2406:cf00:1000::/43
2406:cf00::/48
2406:d440:100::/44
2406:d440:200::/44
2406:d440:300::/44
2406:d440:ff00::/48
2406:d440:ffff::/48
2406:e3c0::/32
2406:e500::/33
2407:23c0::/48
2407:2840::/48
2407:3740::/48
2407:37c0::/32
2407:4980::/32
2407:5380::/32
2407:6c40:1100::/48
2407:6c40:1210::/48
2407:6c40:1500::/48
2407:6c40:1600::/40
2407:8f40:2::/48
2407:9f00::/32
2407:ad80::/32
2407:ae80::/32
2407:b380:8000::/48
2407:bc00::/32
2407:c080:4000::/37
2407:c080:5000::/37
2407:c080:6000::/36
2407:c080:8000::/36
2407:c080::/35
2407:d9c0::/32
2408:4000::/22
2408:8000:1000::/36
2408:8000:2000::/35
2408:8000:2::/47
2408:8000:4000::/34
2408:8000:8000::/33
2408:8000::/48
2408:8001::/32
2408:8020::/30
2408:8024::/31
2408:8026::/32
2408:802a:8000::/33
2408:802c::/32
2408:803e::/32
2408:8056::/32
2408:805c::/30
2408:8060::/33
2408:80c2::/33
2408:80c5::/33
2408:80ca:8000::/34
2408:80ca::/33
2408:80da:8000::/34
2408:80da::/33
2408:80e0:4000::/34
2408:80e0:8000::/33
2408:80e2::/33
2408:80e9:4000::/34
2408:80ea:4000::/34
2408:80ea:8000::/33
2408:80f0:4000::/34
2408:80f0:8000::/33
2408:80f1:100::/43
2408:80f1:120::/44
2408:80f1:160::/43
2408:80f1:180::/43
2408:80f1:1b0::/44
2408:80f1:1c0::/43
2408:80f1:1e0::/44
2408:80f1:200::/40
2408:80f1:40::/43
2408:80f1:70::/44
2408:80f1:80::/41
2408:80f1::/42
2408:80f5:4000::/34
2408:80f9:4000::/34
2408:80fa:4000::/34
2408:80fa:8000::/33
2408:8120:1::/48
2408:8120:2::/48
2408:8120:7000::/36
2408:8140:2000::/48
2408:815f:e000::/35
2408:8181:6000::/40
2408:8181:8000::/40
2408:8181:a000::/40
2408:8181:a220::/44
2408:8181:e000::/40
2408:8182:6000::/40
2408:8182:c000::/40
2408:8183:4000::/40
2408:8183:8000::/40
2408:81a2:2000::/35
2408:81a2:4000::/35
2408:81a3:6000::/35
2408:81a3:c800::/48
2408:81a3:ca66::/48
2408:8206::/31
2408:8208::/29
2408:8210::/30
2408:8214::/31
2408:821a::/31
2408:8220::/31
2408:8226::/32
2408:822a::/31
2408:822e::/31
2408:8230::/29
2408:8238::/31
2408:823c::/31
2408:8240::/32
2408:8244::/30
2408:8248::/30
2408:824c::/32
2408:824e::/31
2408:8250::/30
2408:8254::/32
2408:8256::/31
2408:8258::/30
2408:825c::/31
2408:825f::/32
2408:8260::/32
2408:8262::/31
2408:8264::/31
2408:8266::/32
2408:826a::/32
2408:826c::/30
2408:8270::/32
2408:8274::/30
2408:8278::/31
2408:827a::/32
2408:8306::/31
2408:8308::/30
2408:8310::/30
2408:832a::/32
2408:832e::/31
2408:8330::/30
2408:8338::/32
2408:8340::/32
2408:8344::/30
2408:8348::/30
2408:834e::/31
2408:8350::/30
2408:8354::/32
2408:8356::/31
2408:8358::/30
2408:8360::/30
2408:8364::/31
2408:836c::/30
2408:8374::/30
2408:8378::/31
2408:837a::/32
2408:8406:100::/41
2408:8406:1800::/42
2408:8406:180::/42
2408:8406:1840::/43
2408:8406:18c0::/42
2408:8406:1900::/41
2408:8406:1980::/42
2408:8406:2400::/42
2408:8406:2440::/43
2408:8406:24c0::/42
2408:8406:2500::/41
2408:8406:2580::/42
2408:8406:3000::/42
2408:8406:3040::/43
2408:8406:30c0::/42
2408:8406:3100::/41
2408:8406:3180::/42
2408:8406:3c00::/42
2408:8406:3c40::/43
2408:8406:3cc0::/42
2408:8406:3d00::/41
2408:8406:3d80::/42
2408:8406:40::/43
2408:8406:4800::/42
2408:8406:4840::/43
2408:8406:48c0::/42
2408:8406:4900::/41
2408:8406:4980::/42
2408:8406:5400::/42
2408:8406:5440::/43
2408:8406:54c0::/42
2408:8406:5500::/41
2408:8406:5580::/42
2408:8406:6000::/42
2408:8406:6040::/43
2408:8406:60c0::/42
2408:8406:6100::/41
2408:8406:6180::/42
2408:8406:6c00::/42
2408:8406:6c40::/43
2408:8406:6cc0::/42
2408:8406:6d00::/41
2408:8406:6d80::/42
2408:8406:7800::/42
2408:8406:7840::/43
2408:8406:78c0::/42
2408:8406:7900::/41
2408:8406:7980::/42
2408:8406:8400::/42
2408:8406:8440::/43
2408:8406:84c0::/42
2408:8406:8500::/41
2408:8406:8580::/42
2408:8406:9000::/42
2408:8406:9040::/43
2408:8406:90c0::/42
2408:8406:9100::/41
2408:8406:9180::/42
2408:8406:9c00::/42
2408:8406:9c40::/43
2408:8406:9cc0::/42
2408:8406:9d00::/41
2408:8406:9d80::/42
2408:8406::/42
2408:8406:a800::/42
2408:8406:a840::/43
2408:8406:a8c0::/42
2408:8406:a900::/41
2408:8406:a980::/42
2408:8406:b400::/42
2408:8406:b440::/43
2408:8406:b4c0::/42
2408:8406:b500::/41
2408:8406:b580::/42
2408:8406:c00::/42
2408:8406:c0::/42
2408:8406:c40::/43
2408:8406:cc0::/42
2408:8406:d00::/41
2408:8406:d80::/42
2408:8407:500::/43
2408:8409:100::/41
2408:8409:1800::/40
2408:8409:180::/42
2408:8409:1900::/41
2408:8409:1980::/42
2408:8409:19c0::/43
2408:8409:1c0::/43
2408:8409:2400::/40
2408:8409:2500::/41
2408:8409:2580::/42
2408:8409:25c0::/43
2408:8409:3000::/40
2408:8409:3100::/41
2408:8409:3180::/42
2408:8409:31c0::/43
2408:8409:3c00::/40
2408:8409:3d00::/41
2408:8409:3d80::/42
2408:8409:3dc0::/43
2408:8409:4800::/40
2408:8409:4900::/41
2408:8409:4980::/42
2408:8409:49c0::/43
2408:8409:5400::/40
2408:8409:5500::/41
2408:8409:5580::/42
2408:8409:55c0::/43
2408:8409:6000::/40
2408:8409:6100::/41
2408:8409:6180::/42
2408:8409:61c0::/43
2408:8409:6c00::/40
2408:8409:6d00::/41
2408:8409:6d80::/42
2408:8409:6dc0::/43
2408:8409:7800::/40
2408:8409:7900::/41
2408:8409:7980::/42
2408:8409:79c0::/43
2408:8409:8400::/40
2408:8409:8500::/41
2408:8409:8580::/42
2408:8409:85c0::/43
2408:8409:9000::/40
2408:8409:9100::/41
2408:8409:9180::/42
2408:8409:91c0::/43
2408:8409:9c00::/40
2408:8409:9d00::/41
2408:8409:9d80::/42
2408:8409:9dc0::/43
2408:8409::/40
2408:8409:a800::/40
2408:8409:a900::/41
2408:8409:a980::/42
2408:8409:a9c0::/43
2408:8409:b400::/40
2408:8409:b500::/41
2408:8409:b580::/42
2408:8409:b5c0::/43
2408:8409:c00::/40
2408:8409:d00::/41
2408:8409:d80::/42
2408:8409:dc0::/43
2408:840c:1100::/40
2408:840c:1a00::/40
2408:840c:1c00::/40
2408:840c:1e00::/40
2408:840c:200::/40
2408:840c:2700::/40
2408:840c:2900::/40
2408:840c:2b00::/40
2408:840c:3400::/40
2408:840c:3600::/40
2408:840c:3800::/40
2408:840c:400::/40
2408:840c:4e00::/40
2408:840c:5000::/40
2408:840c:5200::/40
2408:840c:5b00::/40
2408:840c:5d00::/40
2408:840c:5f00::/40
2408:840c:6800::/40
2408:840c:6a00::/40
2408:840c:6c00::/40
2408:840c:7500::/40
2408:840c:7700::/40
2408:840c:7900::/40
2408:840c:8200::/40
2408:840c:8400::/40
2408:840c:8600::/40
2408:840c:8f00::/40
2408:840c:9100::/40
2408:840c:9300::/40
2408:840c:9c00::/40
2408:840c:9e00::/40
2408:840c::/40
2408:840c:a000::/40
2408:840c:a900::/40
2408:840c:ab00::/40
2408:840c:ad00::/40
2408:840c:b600::/40
2408:840c:b800::/40
2408:840c:ba00::/40
2408:840c:c300::/40
2408:840c:c500::/40
2408:840c:c700::/40
2408:840c:d000::/40
2408:840c:d00::/40
2408:840c:d200::/40
2408:840c:d400::/40
2408:840c:dd00::/40
2408:840c:de00::/39
2408:840c:f00::/40
2408:840d:1100::/42
2408:840d:1300::/42
2408:840d:1a00::/42
2408:840d:1c00::/42
2408:840d:1e00::/42
2408:840d:2000::/42
2408:840d:200::/42
2408:840d:2700::/42
2408:840d:2900::/42
2408:840d:2b00::/42
2408:840d:2d00::/42
2408:840d:3400::/42
2408:840d:3600::/42
2408:840d:3800::/42
2408:840d:3a00::/42
2408:840d:400::/42
2408:840d:4e00::/42
2408:840d:5000::/42
2408:840d:5200::/42
2408:840d:5400::/42
2408:840d:5b00::/42
2408:840d:5d00::/42
2408:840d:5f00::/42
2408:840d:600::/42
2408:840d:6100::/42
2408:840d:6800::/42
2408:840d:6a00::/42
2408:840d:6c00::/42
2408:840d:6e00::/42
2408:840d:7500::/42
2408:840d:7700::/42
2408:840d:7900::/42
2408:840d:7b00::/42
2408:840d:8200::/42
2408:840d:8400::/42
2408:840d:8600::/42
2408:840d:8800::/42
2408:840d:8f00::/42
2408:840d:9100::/42
2408:840d:9300::/42
2408:840d:9500::/42
2408:840d:9c00::/42
2408:840d:9e00::/42
2408:840d::/42
2408:840d:a000::/42
2408:840d:a200::/42
2408:840d:a900::/42
2408:840d:ab00::/42
2408:840d:ad00::/42
2408:840d:af00::/42
2408:840d:b600::/42
2408:840d:b800::/42
2408:840d:ba00::/42
2408:840d:bc00::/42
2408:840d:c300::/42
2408:840d:c500::/42
2408:840d:c700::/42
2408:840d:c900::/42
2408:840d:d000::/42
2408:840d:d00::/42
2408:840d:d200::/42
2408:840d:d400::/42
2408:840d:d600::/42
2408:840d:dd00::/42
2408:840d:de00::/42
2408:840d:f00::/42
2408:840e:dd00::/40
2408:840e:de00::/39
2408:8410::/30
2408:8414::/31
2408:8417::/32
2408:8418::/32
2408:841a::/31
2408:841c::/31
2408:841e::/32
2408:8420::/31
2408:8422::/32
2408:8426::/31
2408:842a::/31
2408:842c::/32
2408:842e::/32
2408:8431::/32
2408:8434::/30
2408:8438::/31
2408:843c::/30
2408:8440::/31
2408:8444::/30
2408:8448::/32
2408:844b::/32
2408:844c::/30
2408:8452::/31
2408:8454::/32
2408:8456::/31
2408:8458::/30
2408:845c::/31
2408:8460::/30
2408:8464::/31
2408:8466::/32
2408:8469::/32
2408:846a::/31
2408:846c::/30
2408:8470::/31
2408:8474::/30
2408:8478::/31
2408:847a::/32
2408:84e1::/32
2408:84e2::/31
2408:84e4::/30
2408:84e9::/32
2408:84eb::/32
2408:84ec::/30
2408:84f0::/28
2408:856c::/31
2408:8606::/31
2408:8608::/29
2408:8610::/30
2408:8614::/31
2408:861a::/31
2408:861c::/32
2408:8620::/31
2408:8624::/31
2408:8626::/32
2408:862a::/31
2408:862d::/32
2408:862e::/31
2408:8630::/29
2408:8638::/31
2408:863c::/31
2408:8640::/32
2408:8642::/32
2408:8644::/30
2408:8648::/31
2408:864c::/32
2408:864e::/31
2408:8650::/30
2408:8656::/31
2408:8658::/30
2408:865c::/31
2408:865f::/32
2408:8660::/32
2408:8662::/31
2408:8664::/31
2408:8666::/32
2408:866a::/31
2408:866c::/30
2408:8670::/32
2408:8674::/30
2408:8678::/31
2408:867a::/32
2408:8706::/31
2408:8708::/29
2408:8710::/30
2408:8719::/32
2408:871a::/31
2408:8720::/30
2408:8726::/32
2408:872b::/32
2408:872f::/32
2408:8730::/30
2408:8734::/31
2408:8736::/32
2408:8738::/32
2408:873c::/31
2408:8740::/32
2408:8742::/32
2408:8744::/30
2408:8748::/29
2408:8752::/32
2408:8756::/31
2408:8758::/30
2408:875c::/32
2408:8760::/32
2408:8762::/31
2408:8764::/31
2408:8766::/32
2408:8768::/32
2408:876a::/32
2408:876c::/30
2408:8770::/32
2408:8772::/31
2408:8774::/32
2408:8776::/31
2408:8778::/31
2408:877a::/32
2408:877c::/30
2408:8806:40::/43
2408:8806::/42
2408:880c::/30
2408:8810::/30
2408:8814::/31
2408:8818::/31
2408:882c::/32
2408:883a::/32
2408:8844::/43
2408:8856::/31
2408:8858::/30
2408:8862::/31
2408:8864::/31
2408:8866::/32
2408:886e::/31
2408:8872::/32
2408:8878::/31
2408:887e::/32
2408:8906:20::/44
2408:8907:9000::/44
2408:890c::/31
2408:8912::/31
2408:8914::/30
2408:891c::/32
2408:8920::/32
2408:8924::/32
2408:892c::/32
2408:8936::/32
2408:893a::/32
2408:8940::/32
2408:8948::/32
2408:894c::/32
2408:894e::/32
2408:8956::/31
2408:8958::/30
2408:8962::/31
2408:8964::/31
2408:8966::/32
2408:896c::/32
2408:896e::/31
2408:8972::/32
2408:8978::/30
2408:897e::/32
2408:8a00:c000::/36
2408:8a00:d000::/37
2408:8a00:e000::/35
2408:8a01::/36
2408:8a02:b110::/44
2408:8a02:b120::/44
2408:8a04:8000::/36
2408:8a04:e000::/40
2408:8a05:6000::/35
2408:8a05:8000::/36
2408:8a06:100::/47
2408:8a06::/47
2408:8a21:4000::/35
2408:8a22:9200::/39
2408:8a22:9400::/38
2408:8a22:9800::/40
2408:8a22:9a00::/39
2408:8a22:9c00::/38
2408:8a22:a000::/37
2408:8a23:4000::/34
2408:8a24:4000::/34
2408:8a26:c000::/34
2408:8a27:4000::/35
2409:2000::/31
2409:2002::/32
2409:27fa::/48
2409:27fa:f000::/48
2409:27fb::/48
2409:27fc::/48
2409:27fe::/33
2409:8000::/20
240a:2000::/40
240a:2001:1000::/36
240a:2001:100::/40
240a:4002:1b00::/40
240a:4010:8000::/33
240a:4020:83a::/48
240a:4020:883a::/48
240a:4021:83a::/48
240a:4021:883a::/48
240a:4084:2000::/35
240a:4088:a000::/35
240a:408c:2000::/35
240a:4090:1000::/39
240a:4090:1200::/40
240a:4090:120::/48
240a:4090:2010::/48
240a:4090:2041::/48
240a:4090:2061::/48
240a:4090:250::/48
240a:4090:3000::/39
240a:4090:3200::/40
240a:4090:5000::/39
240a:4090:50::/48
240a:4090:5200::/40
240a:4090:7000::/39
240a:4090:7200::/40
240a:4093::/35
240a:4094:2000::/35
240a:409c:2000::/35
240a:40a4:2000::/35
240a:40ac:2000::/35
240a:40b0:283a::/48
240a:40b0:483a::/48
240a:40b0:683a::/48
240a:40b0:83a::/48
240a:40c0:8000::/43
240a:40c0:8200::/48
240a:40c0:8240::/48
240a:40c0:a000::/43
240a:40c0:c000::/43
240a:40c0:e000::/43
240a:40c1:2000::/43
240a:40c1:4000::/43
240a:40c1:6000::/43
240a:40c1:8000::/43
240a:40c1::/43
240a:40c1:a000::/43
240a:40c1:c000::/43
240a:40c1:e000::/43
240a:40c2:2000::/43
240a:40c2:4000::/43
240a:40c2:6000::/43
240a:40c2:8000::/43
240a:40c2::/43
240a:40c2:a000::/43
240a:40c2:c000::/43
240a:40c2:e000::/43
240a:40c3:2000::/43
240a:40c3:4000::/43
240a:40c3:6000::/43
240a:40c3:8000::/43
240a:40c3::/43
240a:40c3:c000::/43
240a:40c3:c200::/48
240a:40c3:c240::/48
240a:40c3:e000::/43
240a:40c4:2000::/43
240a:40c4:4000::/43
240a:40c4::/43
240a:4172::/31
240a:41b0::/34
240a:41b8:8800::/40
240a:41b8:a000::/40
240a:41b8:d400::/40
240a:41d6::/31
240a:41f2::/31
240a:420a::/31
240a:4224:9000::/44
240a:4224:a000::/44
240a:4224:d000::/44
240a:4224:e000::/44
240a:4230::/31
240a:4242::/31
240a:4280::/26
240a:42c0::/27
240a:42e0::/28
240a:42f0::/29
240a:42f8::/30
240a:6001::/48
240a:a000::/20
240a:c000::/20
240b:e001::/32
240b:e002::/31
240b:e004::/30
240b:e008::/29
240b:e010::/32
240c:4000::/22
240c:6:3::/48
240c:6::/47
240c:c000::/20
240c:f:1::/48
240d:4000::/21
240e::/20
2602:2e0:ff::/48
2602:f7ee:ee::/48
2602:f92a:a478::/48
2602:f92a:d1ff::/48
2602:f92a:dead::/48
2602:f92a:e100::/44
2602:f93b:400::/38
2602:f9ba:10c::/48
2602:f9ba:a8::/48
2602:feda:182::/47
2602:feda:1bf::/48
2602:feda:1d1::/48
2602:feda:1df::/48
2602:feda:2d0::/44
2602:feda:2f0::/44
2605:9d80:8001::/48
2605:9d80:8011::/48
2605:9d80:8021::/48
2605:9d80:8031::/48
2605:9d80:8041::/48
2605:9d80:8081::/48
2605:9d80:9003::/48
2605:9d80:9013::/48
2605:9d80:9023::/48
2605:9d80:9033::/48
2605:9d80:9042::/48
2605:9d80:9071::/48
2605:9d80:9092::/48
2620:57:4004::/48
2804:1e48:9001::/48
2804:1e48:9002::/48
2a01:f100:100::/48
2a01:f100:1f8::/48
2a01:ffc7:100::/40
2a03:5840:126::/48
2a04:3e00:1002::/48
2a04:f580:8010::/47
2a04:f580:8090::/48
2a04:f580:8210::/47
2a04:f580:8290::/48
2a04:f580:9010::/48
2a04:f580:9012::/47
2a04:f580:9020::/48
2a04:f580:9030::/48
2a04:f580:9040::/48
2a04:f580:9050::/48
2a04:f580:9060::/48
2a04:f580:9070::/48
2a04:f580:9080::/48
2a04:f580:9090::/48
2a04:f580:9210::/48
2a04:f580:9212::/47
2a04:f580:9220::/48
2a04:f580:9230::/48
2a04:f580:9240::/48
2a04:f580:9250::/48
2a04:f580:9260::/48
2a04:f580:9270::/48
2a04:f580:9280::/48
2a04:f580:9290::/48
2a05:1087::/32
2a05:dfc3:ff00::/40
2a06:1180:1000::/48
2a06:1281:8000::/36
2a06:3603::/32
2a06:3604::/30
2a06:9f81:4600::/43
2a06:9f81:4640::/44
2a06:9f81:4660::/44
2a06:a005:1c40::/44
2a06:a005:260::/43
2a06:a005:280::/43
2a06:a005:2a0::/44
2a06:a005:8d0::/44
2a06:a005:9e0::/44
2a06:a005:a13::/48
2a06:a005:e9a::/48
2a09:54c6:3000::/36
2a09:54c6:6000::/35
2a09:54c6:b000::/36
2a09:54c6:c000::/36
2a09:54c6:e000::/36
2a09:b280:ff81::/48
2a09:b280:ff83::/48
2a09:b280:ff84::/47
2a0a:2840::/30
2a0a:2845:aab8::/46
2a0a:2845:d647::/48
2a0a:2846::/48
2a0a:6040:ec00::/40
2a0a:6044:6600::/39
2a0a:6044:7a00::/40
2a0b:2542::/48
2a0b:4b81:1001::/48
2a0b:4e07:b8::/47
2a0c:9a40:84e0::/48
2a0c:b641:571::/48
2a0c:b641:d40::/44
2a0e:97c0:5ef::/48
2a0e:97c0:83f::/48
2a0e:9b00::/29
2a0e:aa01:1fff::/48
2a0e:aa06:440::/48
2a0e:aa06:490::/44
2a0e:aa06:4e0::/44
2a0e:aa06:500::/44
2a0e:aa06:520::/48
2a0e:aa06:525::/48
2a0e:aa06:541::/48
2a0e:aa06::/40
2a0e:aa07:e01b::/48
2a0e:aa07:e021::/48
2a0e:aa07:e025::/48
2a0e:aa07:e030::/48
2a0e:aa07:e035::/48
2a0e:aa07:e039::/48
2a0e:aa07:e044::/48
2a0e:aa07:e0e0::/44
2a0e:aa07:e151::/48
2a0e:aa07:e155::/48
2a0e:aa07:e160::/47
2a0e:aa07:e162::/48
2a0e:aa07:e16a::/48
2a0e:aa07:e1a0::/44
2a0e:aa07:e1e1::/48
2a0e:aa07:e1e2::/47
2a0e:aa07:e1e4::/47
2a0e:aa07:e1e6::/48
2a0e:aa07:e200::/44
2a0e:aa07:e210::/48
2a0e:aa07:e21c::/47
2a0e:aa07:e220::/44
2a0e:aa07:f0d0::/46
2a0e:aa07:f0d4::/47
2a0e:aa07:f0d8::/48
2a0e:aa07:f0de::/47
2a0e:b107:12b::/48
2a0e:b107:14a0::/44
2a0e:b107:178d::/48
2a0e:b107:178e::/48
2a0e:b107:272::/48
2a0e:b107:740::/44
2a0e:b107:c10::/48
2a0e:b107:da0::/44
2a0e:b107:dce::/48
2a0f:5707:ac00::/47
2a0f:7803:e300::/40
2a0f:7803:f5d0::/44
2a0f:7803:f5e0::/43
2a0f:7803:f680::/43
2a0f:7803:f6a0::/44
2a0f:7803:f7c0::/42
2a0f:7803:f800::/43
2a0f:7803:f840::/44
2a0f:7803:fa21::/48
2a0f:7803:fa22::/47
2a0f:7803:fa24::/46
2a0f:7803:faf3::/48
2a0f:7803:fe41::/48
2a0f:7803:fe44::/46
2a0f:7803:fe4e::/48
2a0f:7803:fe81::/48
2a0f:7803:fe82::/48
2a0f:7804:f650::/44
2a0f:7804:f9f0::/44
2a0f:7807::/32
2a0f:7d07::/32
2a0f:85c1:ba5::/48
2a0f:85c1:ca0::/44
2a0f:85c1:ce1::/48
2a0f:85c1:cf1::/48
2a0f:9400:6110::/48
2a0f:9400:7700::/48
2a0f:ac00::/29
2a0f:ea47:fc1d::/48
2a10:2f00:15a::/48
2a10:67c2:2::/48
2a10:ccc0:d00::/46
2a10:ccc0:d0a::/47
2a10:ccc0:d0c::/47
2a10:ccc6:66c4::/48
2a10:ccc6:66c6::/48
2a10:ccc6:66c8::/47
2a10:ccc6:66ca::/48
2a10:ccc6:66cc::/47
2a12:f8c3::/36
2a13:1800:10::/48
2a13:1800:300::/44
2a13:1800:80::/44
2a13:1800::/48
2a13:1801:180::/43
2a13:a5c3:ff10::/44
2a13:a5c3:ff21::/48
2a13:a5c3:ff50::/44
2a13:a5c7:1800::/40
2a13:a5c7:2100::/48
2a13:a5c7:2102::/48
2a13:a5c7:2121::/48
2a13:a5c7:2301::/48
2a13:a5c7:2302::/48
2a13:a5c7:23c0::/42
2a13:a5c7:2600::/40
2a13:a5c7:2801::/48
2a13:a5c7:2803::/48
2a13:a5c7:3108::/48
2a13:a5c7:31a0::/43
2a13:a5c7:3301::/48
2a13:a5c7:3304::/48
2a13:a5c7:3306::/47
2a13:aac4:f000::/44
2a14:4c41::/32
2a14:67c1:20::/44
2a14:67c1:702::/47
2a14:67c1:704::/48
2a14:67c1:70::/47
2a14:67c1:73::/48
2a14:67c1:74::/48
2a14:67c1:a010::/44
2a14:67c1:a020::/48
2a14:67c1:a023::/48
2a14:67c1:a024::/48
2a14:67c1:a02a::/48
2a14:67c1:a02f::/48
2a14:67c1:a040::/47
2a14:67c1:a061::/48
2a14:67c1:a064::/48
2a14:67c1:a090::/45
2a14:67c1:a099::/48
2a14:67c1:a100::/43
2a14:67c1:a125::/48
2a14:67c1:a144::/48
2a14:67c1:a150::/44
2a14:67c1:b000::/48
2a14:67c1:b065::/48
2a14:67c1:b066::/48
2a14:67c1:b068::/47
2a14:67c1:b100::/46
2a14:67c1:b105::/48
2a14:67c1:b107::/48
2a14:67c1:b130::/46
2a14:67c1:b134::/47
2a14:67c1:b140::/48
2a14:67c1:b4a1::/48
2a14:67c1:b4a2::/48
2a14:67c1:b4c0::/45
2a14:67c1:b4d0::/44
2a14:67c1:b4e0::/43
2a14:67c1:b500::/47
2a14:67c1:b549::/48
2a14:67c1:b561::/48
2a14:67c1:b563::/48
2a14:67c1:b566::/48
2a14:67c1:b581::/48
2a14:67c1:b582::/48
2a14:67c1:b588::/47
2a14:67c1:b590::/47
2a14:67c1:b599::/48
2a14:67c5:1900::/40
2a14:7580:740::/44
2a14:7580:750::/47
2a14:7580:9200::/40
2a14:7580:9400::/39
2a14:7580:9600::/46
2a14:7580:960c::/48
2a14:7580:d000::/37
2a14:7580:d800::/39
2a14:7580:da00::/40
2a14:7580:e200::/40
2a14:7580:fa01::/48
2a14:7580:fe00::/40
2a14:7580:fff4::/48
2a14:7580:fff7::/48
2a14:7580:fffa::/48
2a14:7581:3100::/40
2a14:7581:3400::/47
2a14:7581:9010::/44
2a14:7581:b10::/48
2a14:7581:b12::/48
2a14:7581:b15::/48
2a14:7581:b20::/46
2a14:7581:b32::/47
2a14:7581:b44::/48
2a14:7581:b46::/47
2a14:7581:b48::/48
2a14:7581:b4a::/48
2a14:7581:b60::/48
2a14:7581:b62::/47
2a14:7581:b64::/46
2a14:7581:b6c::/47
2a14:7581:b6e::/48
2a14:7581:b82::/47
2a14:7581:ba0::/48
2a14:7581:ba2::/47
2a14:7581:ba4::/48
2a14:7581:bbb::/48
2a14:7581:bc2::/48
2a14:7581:bcd::/48
2a14:7581:bff::/48
2a14:7581:ffb::/48
2a14:7581:ffd::/48
2a14:7583:f201::/48
2a14:7583:f203::/48
2a14:7583:f300::/46
2a14:7583:f304::/48
2a14:7583:f4fe::/48
2a14:7583:f500::/48
2a14:7583:f701::/48
2a14:7583:f702::/47
2a14:7583:f704::/47
2a14:7583:f707::/48
2a14:7583:f708::/48
2a14:7583:f743::/48
2a14:7583:f764::/48
2a14:7584::/36
2a14:7c0:4a01::/48
2c0f:f7a8:8011::/48
2c0f:f7a8:8050::/48
2c0f:f7a8:805f::/48
2c0f:f7a8:8150::/48
2c0f:f7a8:815f::/48
2c0f:f7a8:8211::/48
2c0f:f7a8:9010::/48
2c0f:f7a8:9020::/48
2c0f:f7a8:9041::/48
2c0f:f7a8:9210::/47
2c0f:f7a8:9220::/48
|
294coder/Efficient-MIF | 5,496 | utils/progress_utils.py | from typing import Callable, List, Union
from rich.progress import Progress, TextColumn, BarColumn, TimeElapsedColumn, TimeRemainingColumn, SpinnerColumn
from contextlib import nullcontext
class EasyProgress:
tbar: Progress = None
task_desp_ids: dict[str, int] = {}
@classmethod
def console(cls):
assert cls.tbar is not None, '`tbar` has not initialized'
return cls.tbar.console
@classmethod
def close_all_tasks(cls):
if cls.tbar is not None:
for task_id in cls.tbar.task_ids:
cls.tbar.stop_task(task_id)
# set the task_id all unvisible
cls.tbar.update(task_id, visible=False)
@classmethod
def easy_progress(cls,
task_desciptions: list[str],
task_total: list[int],
tbar_kwargs: dict={},
task_kwargs: list[dict[str, Union[str, int]]]=None,
is_main_process: bool=True,
*,
start_tbar: bool=True,
debug: bool=False) -> tuple[Progress, Union[list[int], int]]:
"""get a rich progress bar
Args:
task_desciptions (list[str]): list of task descriptions of `len(task_desciptions)` tasks
task_total (list[int]): list of length each task
tbar_kwargs (dict, optional): kwargs for progress bar. Defaults to {}.
task_kwargs (list[dict[str, Union[str, int]]], optional): task kwargs for each task. Defaults to None.
is_main_process (bool, optional): if is main process. Defaults to True.
start_tbar (bool, optional): start running progress bar when ini. Defaults to True.
debug (bool, optional): debug mode, set progress bar to be unvisible. Defaults to False.
Returns:
tuple[Progress, Union[list[int], int]]: Progress bar and task ids
"""
def _add_task_ids(tbar: Progress, task_desciptions, task_total, task_kwargs):
task_ids = []
if task_kwargs is None:
task_kwargs = [{'visible': False}] * len(task_desciptions)
for task_desciption, task_total, id_task_kwargs in zip(task_desciptions, task_total, task_kwargs):
if task_desciption in list(EasyProgress.task_desp_ids.keys()):
task_id = EasyProgress.task_desp_ids[task_desciption]
task_ids.append(task_id)
else:
task_id = tbar.add_task(task_desciption, total=task_total, **id_task_kwargs)
task_ids.append(task_id)
EasyProgress.task_desp_ids[task_desciption] = task_id
return task_ids if len(task_ids) > 1 else task_ids[0]
def _new_tbar_with_task_ids(task_desciptions, task_total, task_kwargs):
if is_main_process:
if task_kwargs is not None:
assert len(task_desciptions) == len(task_total) == len(task_kwargs)
else:
assert len(task_desciptions) == len(task_total)
# if (console := tbar_kwargs.pop('console', None)) is not None:
# console._color_system = console._detect_color_system()
# if 'console' in tbar_kwargs:
# tbar_kwargs['console']._color_system = tbar_kwargs['console']._detect_color_system()
tbar = Progress(TextColumn("[progress.description]{task.description}"),
BarColumn(),
TextColumn("[progress.percentage]{task.percentage:>3.0f}%"),
SpinnerColumn(),
TimeRemainingColumn(),
TimeElapsedColumn(),
**tbar_kwargs)
EasyProgress.tbar = tbar
task_ids = _add_task_ids(tbar, task_desciptions, task_total, task_kwargs)
return tbar, task_ids
else:
return nullcontext(), [None] * len(task_desciptions) if len(task_desciptions) > 1 else None
def _cached_tbar_with_new_task_ids(task_desciptions, task_total, task_kwargs):
if is_main_process:
tbar = EasyProgress.tbar
task_ids = []
if task_kwargs is None:
task_kwargs = [{'visible': False}] * len(task_desciptions)
task_ids = _add_task_ids(tbar, task_desciptions, task_total, task_kwargs)
return tbar, task_ids
else:
return nullcontext(), [None] * len(task_desciptions) if len(task_desciptions) > 1 else None
if not debug:
if EasyProgress.tbar is not None:
rets = _cached_tbar_with_new_task_ids(task_desciptions, task_total, task_kwargs)
else:
rets = _new_tbar_with_task_ids(task_desciptions, task_total, task_kwargs)
if start_tbar and is_main_process and not EasyProgress.tbar.live._started:
EasyProgress.tbar.start()
return rets
else:
return nullcontext(), [None] * len(task_desciptions) if len(task_desciptions) > 1 else None
|
2977094657/DidaAPI | 1,578 | frontend/docs/api/pomodoros/focus-timeline.md | # 获取专注记录时间线
获取专注记录的时间线数据,支持分页获取历史记录。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/pomodoros/timeline`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
| 参数名 | 类型 | 必填 | 说明 | 示例 |
|--------|------|------|------|------|
| to | string | 否 | 分页参数:上一页最后一条记录的startTime,用于获取更早的数据 | 2025-04-22T08:43:31.000+0000 |
## 响应格式
### 成功响应
```json
[
{
"id": "string",
"tasks": [
{
"taskId": "string",
"title": "string",
"projectName": "string",
"startTime": "2025-04-22T08:43:31.000+0000",
"endTime": "2025-04-22T09:38:58.000+0000"
}
],
"startTime": "2025-04-22T08:43:31.000+0000",
"endTime": "2025-04-22T09:38:58.000+0000",
"pauseDuration": 0,
"etag": "string",
"type": 1,
"added": false
}
]
```
### 响应字段说明
| 字段 | 类型 | 说明 |
|------|------|------|
| id | string | 专注记录ID |
| tasks | array | 关联的任务列表 |
| tasks[].taskId | string | 任务ID |
| tasks[].title | string | 任务标题 |
| tasks[].projectName | string | 项目名称 |
| tasks[].startTime | string | 任务开始时间 |
| tasks[].endTime | string | 任务结束时间 |
| startTime | string | 专注开始时间 |
| endTime | string | 专注结束时间 |
| pauseDuration | number | 暂停时长(毫秒) |
| etag | string | 版本标识 |
| type | number | 专注类型(0=番茄专注,1=正计时专注) |
| added | boolean | 是否已添加 |
## 分页说明
1. **首次请求**:不传 `to` 参数,获取最新的专注记录(约31条)
2. **获取更多**:使用上一页最后一条记录的 `startTime` 作为 `to` 参数
3. **时间格式**:`to` 参数使用ISO 8601格式,如 `2025-04-22T08:43:31.000+0000`
4. **自动转换**:系统会自动将时间转换为时间戳并调整时区
|
2977094657/DidaAPI | 1,099 | frontend/docs/api/pomodoros/focus-hour-distribution.md | # 获取专注时间按小时分布
获取指定日期范围内按小时分布的专注时间统计。
## 接口信息
- **接口URL**: `https://api.dida365.com/api/v2/pomodoros/statistics/dist/clock/{start_date}/{end_date}`
- **请求方法**: `GET`
- **认证要求**: 需要登录认证
- **所属平台**: 滴答清单
## 请求参数
| 参数名 | 类型 | 必填 | 说明 | 示例 |
|--------|------|------|------|------|
| start_date | string | 是 | 开始日期,格式: YYYYMMDD | 20250601 |
| end_date | string | 是 | 结束日期,格式: YYYYMMDD | 20250630 |
## 响应格式
### 成功响应
```json
{
"0": 120,
"1": 33,
"15": 120,
"16": 72,
"17": 120,
"18": 131,
"19": 135,
"20": 129,
"21": 159,
"22": 120,
"23": 120
}
```
### 响应字段说明
| 字段 | 类型 | 说明 |
|------|------|------|
| {hour} | number | 指定小时的专注时长(分钟),hour为24小时制(0-23) |
### 数据说明
- **时间格式**: 使用24小时制,0表示00:00-01:00,23表示23:00-24:00
- **时长单位**: 分钟
- **统计范围**: 指定日期范围内的总计专注时间
- **数据聚合**: 将整个时间段内相同小时的专注时间累加
## 使用说明
1. 确保已完成认证获取会话
2. 提供有效的日期范围参数
3. 日期格式必须为 YYYYMMDD
4. 返回数据为对象格式,键为小时数(字符串),值为专注时长(数字)
5. 只有有专注记录的小时才会出现在响应中
6. 可用于生成24小时专注时间分布图表
## 应用场景
- **时间管理分析**: 了解自己在一天中哪些时段最专注
- **工作习惯优化**: 根据专注时间分布调整工作安排
- **效率统计**: 分析不同时间段的专注效率
- **数据可视化**: 生成24小时专注热力图或柱状图
|
294coder/Efficient-MIF | 16,764 | utils/loss_utils.py | from functools import partial
import random
from typing import Sequence, Union, TYPE_CHECKING
from einops import reduce
from contextlib import contextmanager
import kornia
from kornia.filters import spatial_gradient
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch import Tensor
from torch.autograd import Variable
import numpy as np
from math import exp
import sys
sys.path.append('./')
from utils.misc import is_main_process, exists, default
from utils.log_utils import easy_logger
if TYPE_CHECKING:
from model.base_model import BaseModel
logger = easy_logger()
class PerceptualLoss(nn.Module):
def __init__(self, percep_net="vgg", norm=True):
super(PerceptualLoss, self).__init__()
self.norm = norm
self.lpips_loss = lpips.LPIPS(net=percep_net).cuda()
def forward(self, x, y):
# assert x.shape == y.shape
loss = self.lpips_loss(x, y, normalize=self.norm)
return torch.squeeze(loss).mean()
def gaussian(window_size, sigma):
gauss = torch.Tensor(
[
exp(-((x - window_size // 2) ** 2) / float(2 * sigma**2))
for x in range(window_size)
]
)
return gauss / gauss.sum()
class MaxGradientLoss(torch.nn.Module):
def __init__(self, mean_batch=True) -> None:
super().__init__()
self.register_buffer(
"x_sobel_kernel",
torch.FloatTensor([[-1, 0, 1], [-2, 0, 2], [-1, 0, 1]]).expand(1, 1, 3, 3),
)
self.register_buffer(
"y_sobel_kernel",
torch.FloatTensor([[-1, -2, -1], [0, 0, 0], [1, 2, 1]]).expand(1, 1, 3, 3),
)
self.mean_batch = mean_batch
def forward(self, fuse, ir, vis):
c = fuse.size(1)
fuse_grad_x = F.conv2d(fuse, self.x_sobel_kernel, padding=1, groups=c)
fuse_grad_y = F.conv2d(fuse, self.y_sobel_kernel, padding=1, groups=c)
ir_grad_x = F.conv2d(ir, self.x_sobel_kernel, padding=1, groups=c)
ir_grad_y = F.conv2d(ir, self.y_sobel_kernel, padding=1, groups=c)
vis_grad_x = F.conv2d(vis, self.x_sobel_kernel, padding=1, groups=c)
vis_grad_y = F.conv2d(vis, self.y_sobel_kernel, padding=1, groups=c)
max_grad_x = torch.maximum(ir_grad_x, vis_grad_x)
max_grad_y = torch.maximum(ir_grad_y, vis_grad_y)
if self.mean_batch:
max_gradient_loss = (
F.l1_loss(fuse_grad_x, max_grad_x) + F.l1_loss(fuse_grad_y, max_grad_y)
) / 2
else:
x_loss_b = F.l1_loss(fuse_grad_x, max_grad_x, reduction="none").mean(
dim=(1, 2, 3)
)
y_loss_b = F.l1_loss(fuse_grad_y, max_grad_y, reduction="none").mean(
dim=(1, 2, 3)
)
max_gradient_loss = (x_loss_b + y_loss_b) / 2
return max_gradient_loss
def create_window(window_size, channel, sigma=1.5):
_1D_window = gaussian(window_size, sigma).unsqueeze(1)
_2D_window = _1D_window.mm(_1D_window.t()).float().unsqueeze(0).unsqueeze(0)
window = Variable(
_2D_window.expand(channel, 1, window_size, window_size).contiguous()
)
return window
def mci_loss(pred, gt):
return F.l1_loss(pred, gt.max(1, keepdim=True)[0])
def sf(f1, kernel_radius=5):
"""copy from https://github.com/tthinking/YDTR/blob/main/losses/__init__.py
Args:
f1 (torch.Tensor): image shape [b, c, h, w]
kernel_radius (int, optional): kernel redius using calculate sf. Defaults to 5.
Returns:
loss: loss item. type torch.Tensor
"""
device = f1.device
b, c, h, w = f1.shape
r_shift_kernel = (
torch.FloatTensor([[0, 0, 0], [1, 0, 0], [0, 0, 0]])
.to(device)
.reshape((1, 1, 3, 3))
.repeat(c, 1, 1, 1)
)
b_shift_kernel = (
torch.FloatTensor([[0, 1, 0], [0, 0, 0], [0, 0, 0]])
.to(device)
.reshape((1, 1, 3, 3))
.repeat(c, 1, 1, 1)
)
f1_r_shift = F.conv2d(f1, r_shift_kernel, padding=1, groups=c)
f1_b_shift = F.conv2d(f1, b_shift_kernel, padding=1, groups=c)
f1_grad = torch.pow((f1_r_shift - f1), 2) + torch.pow((f1_b_shift - f1), 2)
kernel_size = kernel_radius * 2 + 1
add_kernel = torch.ones((c, 1, kernel_size, kernel_size)).float().to(device)
kernel_padding = kernel_size // 2
f1_sf = torch.sum(
F.conv2d(f1_grad, add_kernel, padding=kernel_padding, groups=c), dim=1
)
return 1 - f1_sf
def _ssim(img1, img2, window, window_size, channel, size_average=True):
mu1 = F.conv2d(img1, window, padding=window_size // 2, groups=channel)
mu2 = F.conv2d(img2, window, padding=window_size // 2, groups=channel)
mu1_sq = mu1.pow(2)
mu2_sq = mu2.pow(2)
mu1_mu2 = mu1 * mu2
sigma1_sq = (
F.conv2d(img1 * img1, window, padding=window_size // 2, groups=channel) - mu1_sq
)
sigma2_sq = (
F.conv2d(img2 * img2, window, padding=window_size // 2, groups=channel) - mu2_sq
)
sigma12 = (
F.conv2d(img1 * img2, window, padding=window_size // 2, groups=channel)
- mu1_mu2
)
C1 = 0.01**2
C2 = 0.03**2
ssim_map = ((2 * mu1_mu2 + C1) * (2 * sigma12 + C2)) / (
(mu1_sq + mu2_sq + C1) * (sigma1_sq + sigma2_sq + C2)
)
if size_average:
return ssim_map.mean()
else:
return ssim_map.mean(1).mean(1).mean(1)
class HybridL1L2(torch.nn.Module):
def __init__(self):
super(HybridL1L2, self).__init__()
self.l1 = torch.nn.L1Loss()
self.l2 = torch.nn.MSELoss()
self.loss = LossWarpper(l1=self.l1, l2=self.l2)
def forward(self, pred, gt):
loss, loss_dict = self.loss(pred, gt)
return loss, loss_dict
class HybridSSIMSF(torch.nn.Module):
def __init__(self, channel, weighted_r=(1.0, 5e-2, 6e-4, 25e-5)) -> None:
super().__init__()
self.weighted_r = weighted_r
def forward(self, fuse, gt):
# fuse: [b, 1, h, w]
vi = gt[:, 0:1] # [b, 1, h, w]
ir = gt[:, 1:] # [b, 1, h, w]
_ssim_f_ir = ssim_loss_ir(fuse, ir)
_ssim_f_vi = ssim_loss_vi(fuse, vi)
_sf_f_ir = sf_loss_ir(fuse, ir)
_sf_f_vi = sf_loss_vi(fuse, vi)
ssim_f_ir = self.weighted_r[0] * _ssim_f_ir
ssim_f_vi = self.weighted_r[1] * _ssim_f_vi
sf_f_ir = self.weighted_r[2] * _sf_f_ir
sf_f_vi = self.weighted_r[3] * _sf_f_vi
loss_dict = dict(
ssim_f_ir=ssim_f_ir,
ssim_f_vi=ssim_f_vi,
sf_f_ir=sf_f_ir,
sf_f_vi=sf_f_vi,
)
loss = ssim_f_ir + ssim_f_vi + sf_f_ir + sf_f_vi
return loss, loss_dict
class HybridSSIMMCI(torch.nn.Module):
def __init__(self, channel, weight_r=(1.0, 1.0, 1.0)) -> None:
super().__init__()
self.ssim = SSIMLoss(channel=channel)
self.mci_loss = mci_loss
self.weight_r = weight_r
def forward(self, fuse, gt):
# fuse: [b, 1, h, w]
vi = gt[:, 0:1] # [b, 1, h, w]
ir = gt[:, 1:] # [b, 1, h, w]
_ssim_f_ir = self.weight_r[0] * self.ssim(fuse, ir)
_ssim_f_vi = self.weight_r[1] * self.ssim(fuse, vi)
_mci_loss = self.weight_r[2] * self.mci_loss(fuse, gt)
loss = _ssim_f_ir + _ssim_f_vi + _mci_loss
loss_dict = dict(
ssim_f_ir=_ssim_f_ir,
ssim_f_vi=_ssim_f_vi,
mci_loss=_mci_loss,
)
return loss, loss_dict
def accum_loss_dict(ep_loss_dict: dict, loss_dict: dict):
for k, v in loss_dict.items():
if k in ep_loss_dict:
ep_loss_dict[k] += v
else:
ep_loss_dict[k] = v
return ep_loss_dict
def ave_ep_loss(ep_loss_dict: dict, ep_iters: int):
for k, v in ep_loss_dict.items():
ep_loss_dict[k] = v / ep_iters
return ep_loss_dict
@is_main_process
def ave_multi_rank_dict(rank_loss_dict: "list[dict] | dict"):
# type is dict is only one process
assert isinstance(rank_loss_dict, (list, dict)), 'rank_loss_dict must be a list or a dict'
if isinstance(rank_loss_dict, dict):
return rank_loss_dict
n = len(rank_loss_dict)
if n == 1:
return rank_loss_dict[0]
ave_dict = {}
keys = rank_loss_dict[0].keys()
for k in keys:
vs = 0
for d in rank_loss_dict:
v = d[k]
vs = vs + v
ave_dict[k] = vs / n
return ave_dict
class HybridL1SSIM(torch.nn.Module):
def __init__(self, channel=31, weighted_r=(1.0, 0.1)):
super(HybridL1SSIM, self).__init__()
assert len(weighted_r) == 2
self._l1 = torch.nn.L1Loss()
self._ssim = SSIMLoss(channel=channel)
self.loss = LossWarpper(weighted_r, l1=self._l1, ssim=self._ssim)
def forward(self, pred, gt):
loss, loss_dict = self.loss(pred, gt)
return loss, loss_dict
class HybridCharbonnierSSIM(torch.nn.Module):
def __init__(self, weighted_r, channel=31) -> None:
super().__init__()
self._ssim = SSIMLoss(channel=channel)
self._charb = CharbonnierLoss(eps=1e-4)
self.loss = LossWarpper(weighted_r, charbonnier=self._charb, ssim=self._ssim)
def forward(self, pred, gt):
loss, loss_dict = self.loss(pred, gt)
return (loss,)
class HybridMCGMCI(torch.nn.Module):
def __init__(self, weight_r=(1.0, 1.0)) -> None:
super().__init__()
self.mcg = MaxGradientLoss()
self.mci = mci_loss
self.weight_r = weight_r
def forward(self, pred, gt):
vis = gt[:, 0:1]
ir = gt[:, 1:]
mcg_loss = self.mcg(pred, ir, vis) * self.weight_r[0]
mci_loss = self.mci(pred, gt) * self.weight_r[1]
loss_dict = dict(mcg=mcg_loss, mci=mci_loss)
return mcg_loss + mci_loss, loss_dict
def gradient(input):
"""
求图像梯度, sobel算子
:param input:
:return:
"""
filter1 = nn.Conv2d(
kernel_size=3, in_channels=1, out_channels=1, bias=False, padding=1, stride=1
)
filter2 = nn.Conv2d(
kernel_size=3, in_channels=1, out_channels=1, bias=False, padding=1, stride=1
)
filter1.weight.data = (
torch.tensor([[-1.0, 0.0, 1.0], [-2.0, 0.0, 2.0], [-1.0, 0.0, 1.0]])
.reshape(1, 1, 3, 3)
.to(input.device)
)
filter2.weight.data = (
torch.tensor([[1.0, 2.0, 1.0], [0.0, 0.0, 0.0], [-1.0, -2.0, -1.0]])
.reshape(1, 1, 3, 3)
.to(input.device)
)
g1 = filter1(input)
g2 = filter2(input)
image_gradient = torch.abs(g1) + torch.abs(g2)
return image_gradient
class LossWarpper(torch.nn.Module):
def __init__(self, weighted_ratio=(1.0, 1.0), **losses):
super(LossWarpper, self).__init__()
self.names = []
assert len(weighted_ratio) == len(losses.keys())
self.weighted_ratio = weighted_ratio
for k, v in losses.items():
self.names.append(k)
setattr(self, k, v)
def forward(self, pred, gt) -> tuple[torch.Tensor, dict[torch.Tensor]]:
loss = 0.0
d_loss = {}
for i, n in enumerate(self.names):
l = getattr(self, n)(pred, gt) * self.weighted_ratio[i]
loss += l
d_loss[n] = l
return loss, d_loss
class TorchLossWrapper(torch.nn.Module):
def __init__(self, weight_ratio: Union[tuple[float], list[float]], **loss) -> None:
super().__init__()
self.key = list(loss.keys())
self.loss = list(loss.values())
self.weight_ratio = weight_ratio
assert len(weight_ratio) == len(loss.keys())
def forward(self, pred, gt):
loss_total = 0.0
loss_d = {}
for i, l in enumerate(self.loss):
loss_i = l(pred, gt) * self.weight_ratio[i]
loss_total = loss_total + loss_i
k = self.key[i]
loss_d[k] = loss_i
return loss_total, loss_d
class SSIMLoss(torch.nn.Module):
def __init__(
self, win_size=11, win_sigma=1.5, data_range=1, size_average=True, channel=3
):
super(SSIMLoss, self).__init__()
self.window_size = win_size
self.size_average = size_average
self.channel = channel
self.window = create_window(win_size, self.channel, win_sigma)
self.win_sigma = win_sigma
def forward(self, img1, img2):
# print(img1.size())
(_, channel, _, _) = img1.size()
if channel == self.channel and self.window.data.type() == img1.data.type():
window = self.window
else:
window = create_window(self.window_size, channel, self.win_sigma)
if img1.is_cuda:
window = window.cuda(img1.get_device())
window = window.type_as(img1)
self.window = window
self.channel = channel
return 1 - _ssim(
img1, img2, window, self.window_size, channel, self.size_average
)
def ssim(img1, img2, win_size=11, data_range=1, size_average=True):
(_, channel, _, _) = img1.size()
window = create_window(win_size, channel)
if img1.is_cuda:
window = window.cuda(img1.get_device())
window = window.type_as(img1)
return _ssim(img1, img2, window, win_size, channel, size_average)
def elementwise_charbonnier_loss(
input: Tensor, target: Tensor, eps: float = 1e-3
) -> Tensor:
"""Apply element-wise weight and reduce loss between a pair of input and
target.
"""
return torch.sqrt((input - target) ** 2 + (eps * eps))
class HybridL1L2(nn.Module):
def __init__(self, cof=10.0):
super(HybridL1L2, self).__init__()
self.l1 = nn.L1Loss()
self.l2 = nn.MSELoss()
self.cof = cof
def forward(self, pred, gt):
return self.l1(pred, gt) / self.cof + self.l2(pred, gt)
def get_loss(loss_type, channel=31, **kwargs: "Unpack[GetLossKwargsdType]"):
if loss_type == "mse":
criterion = nn.MSELoss()
elif loss_type == "l1":
criterion = TorchLossWrapper((1.,), l1=nn.L1Loss())
elif loss_type == "hybrid":
criterion = HybridL1L2()
elif loss_type == "smoothl1":
criterion = nn.SmoothL1Loss()
elif loss_type == "l1ssim":
criterion = HybridL1SSIM(channel=channel, weighted_r=(1.0, 0.1))
else:
raise NotImplementedError(f"loss {loss_type} is not implemented")
return criterion
if __name__ == "__main__":
# loss = SSIMLoss(channel=31)
# loss = CharbonnierLoss(eps=1e-3)
# x = torch.randn(1, 31, 64, 64, requires_grad=True)
# y = x + torch.randn(1, 31, 64, 64) / 10
# l = loss(x, y)
# l.backward()
# print(l)
# print(x.grad)
# import PIL.Image as Image
# vi = (
# np.array(
# Image.open(
# "/media/office-401/Elements SE/cao/ZiHanCao/datasets/RoadScene_and_TNO/training_data/vi/FLIR_05857.jpg"
# ).convert("L")
# )
# / 255
# )
# ir = (
# np.array(
# Image.open(
# "/media/office-401/Elements SE/cao/ZiHanCao/datasets/RoadScene_and_TNO/training_data/ir/FLIR_05857.jpg"
# ).convert("L")
# )
# / 255
# )
# torch.cuda.set_device("cuda:0")
# vi = torch.tensor(vi)[None, None].float() # .cuda()
# ir = torch.tensor(ir)[None, None].float() # .cuda()
# fuse = ((vi + ir) / 2).repeat_interleave(2, dim=0)
# fuse.requires_grad_()
# print(fuse.requires_grad)
# gt = torch.cat((vi, ir), dim=1).repeat_interleave(2, dim=0)
# fuse_loss = HybridSSIMRMIFuse(weight_ratio=(1.0, 1.0, 1.0), ssim_channel=1)
torch.cuda.set_device("cuda:1")
class FuseModel:
def only_fusion_step(self, a, b):
return a + b
fuse_loss = DRMFFusionLoss(grad_op='sobel', reduce_label=True, pseudo_l1_const=0.).cuda()
# fuse_loss = get_emma_fusion_loss(FuseModel())
# print(fuse_loss(fused, (vis, ir), mask))
# u2fusion_loss = U2FusionLoss().cuda()
import time
while True:
fused = torch.randn(1, 3, 64, 64).cuda().requires_grad_()
vis = torch.randn(1, 3, 64, 64).cuda().requires_grad_()
ir = torch.randn(1, 3, 64, 64).cuda().requires_grad_()
mask = torch.randint(0, 3, (1, 1, 64, 64)).cuda().float()
loss = fuse_loss(fused, (vis, ir), mask=mask)
loss[0].backward()
print(loss)
time.sleep(0.1)
# fuse_loss = HybridPIALoss().cuda(1)
# fuse_loss = CDDFusionLoss() # .cuda()
# loss, loss_d = fuse_loss(fuse, gt)
# loss.backward()
# print(loss)
# print(loss_d)
# print(fuse.grad)
# mcg_mci_loss = HybridMCGMCI()
# print(mcg_mci_loss(fuse, gt))
|
281677160/openwrt-package | 79,959 | luci-app-passwall/root/usr/share/passwall/rules/gfwlist | 000webhost.com
030buy.com
0rz.tw
1000giri.net
10beasts.net
10conditionsoflove.com
10musume.com
123rf.com
12bet.com
12vpn.com
12vpn.net
1337x.to
138.com
141hongkong.com
141jj.com
141tube.com
1688.com.au
173ng.com
177pic.info
17t17p.com
18board.com
18comic.org
18onlygirls.com
18p2p.com
18virginsex.com
1984bbs.com
1991way.com
1dumb.com
1e100.net
1eew.com
1lib.sk
1mobile.com
1point3acres.com
1pondo.tv
2-hand.info
2000fun.com
2008xianzhang.info
2021hkcharter.com
2047.name
2047.one
2049bbs.xyz
21andy.com
21sextury.com
228.net.tw
233abc.com
24hrs.ca
25u.com
2lipstube.com
2shared.com
2waky.com
3-a.net
30boxes.com
315lz.com
32red.com
36rain.com
3a5a.com
3arabtv.com
3boys2girls.com
3d-game.com
3proxy.ru
3ren.ca
3tui.net
404museum.com
466453.com
4bluestones.biz
4chan.com
4dq.com
4everproxy.com
4irc.com
4mydomain.com
4pu.com
4rbtv.com
4shared.com
4sqi.net
500px.com
500px.org
51.ca
51jav.org
51luoben.com
5278.cc
5299.tv
56cun04.jigsy.com
5i01.com
5isotoi5.org
5maodang.com
611study.com
611study.icu
63i.com
64museum.org
64tianwang.com
64wiki.com
66.ca
666kb.com
666pool.cn
69shuba.cx
6do.news
6do.world
6park.com
6parkbbs.com
6parker.com
6parknews.com
7capture.com
7cow.com
8-d.com
85cc.us
881903.com
888.com
888poker.com
89-64.org
89.64.charter.constitutionalism.solutions
8964museum.com
8news.com.tw
8z1.net
91dasai.com
91porn.com
91porny.com
91vps.club
92ccav.com
991.com
99btgc01.com
99cn.info
9bis.com
9bis.net
9cache.com
9gag.com
9news.com.au
a-normal-day.com
a248.e.akamai.net
a5.com.ru
aamacau.com
abc.com
abc.net.au
abc.xyz
abchinese.com
abebooks.co.uk
abebooks.com
abematv.akamaized.net
abitno.linpie.com
ablwang.com
aboluowang.com
about.gitlab.com
about.me
abplive.com
abs.edu
acast.com
accim.org
accountkit.com
aceros-de-hispania.com
acevpn.com
acg.rip
acg18.me
acgbox.org
acgkj.com
acgnx.se
acmedia365.com
acmetoy.com
acnw.com.au
actfortibet.org
actimes.com.au
activpn.com
aculo.us
adcex.com
addictedtocoffee.de
addons.mozilla.org
addyoutube.com
adelaidebbs.com
admin.recaptcha.net
admob.com
adpl.org.hk
ads-twitter.com
adsense.com
adult-sex-games.com
adult.friendfinder.com
adultfriendfinder.com
advanscene.com
advertfan.com
advertisercommunity.com
ae.hao123.com
ae.org
aei.org
aenhancers.com
aex.com
af.mil
afantibbs.com
afr.com
afreecatv.com
agnesb.fr
agoogleaday.com
agro.hk
ai.binwang.me
aiosearch.com
aiph.net
airasia.com
airconsole.com
airitilibrary.com
airvpn.org
aisex.com
aiss.anws.gov.tw
ait.org.tw
aiweiwei.com
aiweiweiblog.com
akademiye.org
akamaihd.net
akiba-online.com
akiba-web.com
akinator.com
akow.org
al-islam.com
alabout.com
alanhou.com
alarab.qa
alasbarricadas.org
alforattv.net
alhayat.com
alicejapan.co.jp
aliengu.com
alive.bar
aljazeera.com
alkasir.com
all4mom.org
allcoin.com
allconnected.co
alldrawnsex.com
allfinegirls.com
allgirlmassage.com
allgirlsallowed.org
allgravure.com
alliance.org.hk
allinfa.com
alljackpotscasino.com
allmovie.com
allowed.org
almostmy.com
alphaporno.com
alternate-tools.com
alternativeto.net
altrec.com
alvinalexander.com
alwaysdata.com
alwaysdata.net
alwaysvpn.com
am730.com.hk
amazon.co.jp
amazonvideo.com
ameblo.jp
americangreencard.com
americanunfinished.com
americorps.gov
amiblockedornot.com
amigobbs.net
amitabhafoundation.us
amnesty.org
amnesty.org.hk
amnesty.tw
amnestyusa.org
ampproject.org
amtb-taipei.org
amuletmc.com
anchor.fm
anchorfree.com
ancsconf.org
andfaraway.net
android-x86.org
android.com
androidapksfree.com
androidify.com
androidtv.com
andygod.com
angela-merkel.de
angelfire.com
angola.org
angularjs.org
animecrazy.net
aniscartujo.com
annas-archive.org
annas-archive.se
annatam.com
anobii.com
anonfiles.com
anontext.com
anonymitynetwork.com
anonymizer.com
anonymouse.org
anpopo.com
answering-islam.org
anthonycalzadilla.com
anthropic.com
antichristendom.com
antiwave.net
antpool.com
anyporn.com
anysex.com
ao3.org
aobo.com.au
aofriend.com
aojiao.org
aomedia.org
aomiwang.com
apartmentratings.com
apartments.com
apat1989.org
apetube.com
api-secure.recaptcha.net
api-verify.recaptcha.net
api.ai
api.pureapk.com
api.recaptcha.net
api.steampowered.com
apiary.io
apigee.com
apk.support
apkcombo.com
apkmirror.com
apkmonk.com
apkplz.com
apkpure.com
apkpure.net
app.box.com
app.cloudcone.com
app.smartmailcloud.com
appadvice.com
appbrain.com
appdownloader.net
appledaily.com
appledaily.com.tw
apps.evozi.com
appshopper.com
appsocks.net
appspot.com
appsto.re
aptoide.com
ar.hao123.com
archive.fo
archive.is
archive.li
archive.md
archive.org
archive.ph
archive.today
archive.vn
archiveofourown.com
archiveofourown.org
archives.gov
archives.gov.tw
arctosia.com
areca-backup.org
arena.taipei
arethusa.su
arlingtoncemetery.mil
art4tibet1998.org
arte.tv
artofpeacefoundation.org
artstation.com
artsy.net
arvanstorage.ir
asacp.org
asdfg.jp
asg.to
asia-gaming.com
asiaharvest.org
asianage.com
asianews.it
asiansexdiary.com
asiaone.com
asiatgp.com
ask.com
askstudent.com
askynz.net
aspi.org.au
aspistrategist.org.au
assembla.com
assets.bwbx.io
assimp.org
astrill.com
atc.org.au
atchinese.com
atgfw.org
athenaeizou.com
atlanta168.com
atnext.com
auctions.yahoo.co.jp
audacy.com
auntology.fandom.com
authorizeddns.net
authorizeddns.org
autodraw.com
av-e-body.com
av.com
av.movie
av01.tv
avaaz.org
avcool.com
avdb.in
avdb.tv
avfantasy.com
avg.com
avgle.com
avidemux.org
avmo.pw
avmoo.com
avmoo.net
avmoo.pw
avoision.com
avyahoo.com
axios.com
axureformac.com
azerimix.com
azirevpn.com
azurewebsites.net
b-ok.cc
b.hatena.ne.jp
b0ne.com
babylonbee.com
babynet.com.hk
backchina.com
backpackers.com.tw
backtotiananmen.com
bad.news
badiucao.com
badjojo.com
badoo.com
bahamut.com.tw
baidu.jp
baijie.org
bailandaily.com
baixing.me
baizhi.org
banana-vpn.com
band.us
bandcamp.com
bandwagonhost.com
bangbrosnetwork.com
bangchen.net
bangdream.space
bangkokpost.com
bangumi.moe
bangyoulater.com
bankmobilevibe.com
bannedbook.org
bannednews.org
banorte.com
baramangaonline.com
barenakedislam.com
barnabu.co.uk
bartender.dowjones.com
barton.de
bastillepost.com
bayvoice.net
bb-chat.tv
bbc.co.uk
bbc.com
bbc.in
bbcchinese.com
bbchat.tv
bbci.co.uk
bbg.gov
bbkz.com
bbnradio.org
bbs-tw.com
bbs.brockbbs.com
bbs.cantonese.asia
bbs.ecstart.com
bbs.hanminzu.org
bbs.huasing.org
bbs.junglobal.net
bbs.mikocon.com
bbs.morbell.com
bbs.mychat.to
bbs.naixi.net
bbs.nyinfor.com
bbs.sina.com
bbs.skykiwi.com
bbs.sou-tong.org
bbsdigest.com
bbsland.com
bbsmo.com
bbsone.com
bbtoystore.com
bcc.com.tw
bcchinese.net
bcex.ca
bcmorning.com
bdsmvideos.net
beaconevents.com
bearteach.com
bebo.com
beeg.com
beepool.com
beepool.org
beevpn.com
behance.net
behindkink.com
beijing1989.com
beijing2022.art
beijingspring.com
belamionline.com
bell.wiki
bemywife.cc
beric.me
berlinerbericht.de
berlintwitterwall.com
berm.co.nz
bestgore.com
bestpornstardb.com
bestvpn.com
bestvpnanalysis.com
bestvpnforchina.net
bestvpnserver.com
bestvpnservice.com
bestvpnusa.com
bet365.com
betaclouds.net
betfair.com
betterhash.net
betternet.co
bettervpn.com
bettween.com
betvictor.com
bewww.net
beyondfirewall.com
bfnn.org
bfsh.hk
bgme.me
bgvpn.com
bianlei.com
biantailajiao.com
biblesforamerica.org
bibox.com
biedian.me
big.one
bigfools.com
bigjapanesesex.com
bigmoney.biz
bignews.org
bigone.com
bigsound.org
bild.de
biliworld.com
billypan.com
binance.com
binance.org
binancezh.cc
binux.me
bird.so
bit-z.com
bit.do
bit.ly
bitbay.net
bitchute.com
bitcointalk.org
bitcoinworld.com
bitfinex.com
bitget.com
bithumb.com
bitmex.com
bitshare.com
bitsnoop.com
bitterwinter.org
bitvise.com
bitz.ai
bizhat.com
bjnewlife.org
bjs.org
bjzc.org
bl-doujinsouko.com
blacked.com
blacklogic.com
blackmagicdesign.com
blackvpn.com
blewpass.com
blinkx.com
blinw.com
blip.tv
blockcast.it
blockcn.com
blockedbyhk.com
blockless.com
blocktempo.com
blog.cryptographyengineering.com
blog.de
blog.excite.co.jp
blog.expofutures.com
blog.fizzik.com
blog.foolsmountain.com
blog.fuckgfw233.org
blog.goo.ne.jp
blog.jackjia.com
blog.jp
blog.lester850.info
blog.martinoei.com
blog.pathtosharepoint.com
blog.pentalogic.net
blog.ranxiang.com
blog.reimu.net
blog.sogoo.org
blog.soylent.com
blog.syx86.com
blog.taragana.com
blog.tiney.com
blog.youthwant.com.tw
blogblog.com
blogcatalog.com
blogcity.me
blogdns.org
blogger.com
blogimg.jp
blogjav.net
bloglines.com
bloglovin.com
blogspot.ae
blogspot.al
blogspot.am
blogspot.ba
blogspot.be
blogspot.bg
blogspot.ca
blogspot.cat
blogspot.ch
blogspot.cl
blogspot.co.uk
blogspot.com
blogspot.com.ar
blogspot.com.au
blogspot.com.br
blogspot.com.by
blogspot.com.co
blogspot.com.cy
blogspot.com.ee
blogspot.com.eg
blogspot.com.es
blogspot.com.mt
blogspot.com.ng
blogspot.com.tr
blogspot.com.uy
blogspot.cz
blogspot.de
blogspot.dk
blogspot.fi
blogspot.fr
blogspot.gr
blogspot.hk
blogspot.hr
blogspot.hu
blogspot.ie
blogspot.in
blogspot.is
blogspot.it
blogspot.jp
blogspot.kr
blogspot.li
blogspot.lt
blogspot.lu
blogspot.md
blogspot.mk
blogspot.mx
blogspot.my
blogspot.nl
blogspot.no
blogspot.pe
blogspot.pt
blogspot.qa
blogspot.ro
blogspot.ru
blogspot.se
blogspot.sg
blogspot.si
blogspot.sk
blogspot.sn
blogspot.tw
blogspot.ug
blogtd.org
bloodshed.net
bloomberg.cn
bloomberg.com
bloomberg.de
bloombergview.com
bloomfortune.com
blubrry.com
blueangellive.com
bmdru.com
bnbstatic.com
bnext.com.tw
bnn.co
bnrmetal.com
boardreader.com
bod.asia
bodog88.com
bolehvpn.net
bolin.netfirms.com
bonbonme.com
bonfoundation.org
bongacams.com
boobstagram.com
book.com.tw
book.zi5.me
bookdepository.com
bookepub.com
books.com.tw
booktopia.com.au
bookwalker.com.tw
bootstrapcdn.com
borgenmagazine.com
bot.nu
botanwang.com
bowenpress.com
boxpn.com
boxun.com
boxun.tv
boxunclub.com
boyangu.com
boyfriendtv.com
boysfood.com
boysmaster.com
br.hao123.com
br.st
brainyquote.com
braumeister.org
brave.com
bravotube.net
brazzers.com
breached.to
break.com
breakgfw.com
breaking911.com
breakingtweets.com
breakwall.net
briian.com
brill.com
brizzly.com
broadbook.com
broadpressinc.com
brookings.edu
brucewang.net
brutaltgp.com
bsky.app
bsky.network
bsky.social
bt4g.org
bt4gprx.com
bt95.com
btaia.com
btbit.net
btbtav.com
btbtt.co
btbtt.me
btc.com
btc98.com
btcbank.bank
btctrade.im
btdig.com
btdigg.org
btguard.com
btku.me
btku.org
btspread.com
btsynckeys.com
budaedu.org
buddhanet.com.tw
buffered.com
bullguard.com
bullog.org
bullogger.com
bumingbai.net
bunbunhk.com
busayari.com
business-humanrights.org
business.page
businessinsider.com
businesstoday.com.tw
businessweek.com
busu.org
busytrade.com
buzzhand.com
buzzhand.net
buzzorange.com
buzzsprout.com
bvpn.com
bwgyhw.com
bwh1.net
bx.in.th
bybit.com
bynet.co.il
bypasscensorship.org
byrut.org
c-est-simple.com
c-span.org
c-spanvideo.org
c.mi.com
c100tibet.org
c2cx.com
c3pool.com
cableav.tv
cablegatesearch.net
cachefly.com
cachefly.net
cachinese.com
cacnw.com
cactusvpn.com
cafepress.com
cahr.org.tw
calameo.com
calebelston.com
calendarz.com
calgarychinese.ca
calgarychinese.com
calgarychinese.net
cam4.com
cam4.jp
cam4.sg
camfrog.com
campaign-archive.com
campaignforuyghurs.org
cams.com
cams.org.sg
canadameet.com
canalporno.com
canyu.org
caobian.info
caochangqing.com
caoporn.us
cap.org.hk
carabinasypistolas.com
cardinalkungfoundation.org
carfax.com
cari.com.my
caribbeancom.com
carmotorshow.com
carousell.com.hk
carrd.co
cartoonmovement.com
casadeltibetbcn.org
casatibet.org.mx
casino.williamhill.com
casinobellini.com
casinoking.com
casinoriva.com
castbox.fm
catbox.moe
catch22.net
catchgod.com
catholic.org.hk
catholic.org.tw
cathvoice.org.tw
cato.org
cattt.com
caus.com
cbc.ca
cbs.ntu.edu.tw
cbsnews.com
cbtc.org.hk
cccat.cc
cccat.co
ccfd.org.tw
cchere.com
ccim.org
cclife.ca
cclife.org
cclifefl.org
ccthere.com
ccthere.net
cctmweb.net
cctongbao.com
ccue.ca
ccue.com
ccvoice.ca
ccw.org.tw
cdbook.org
cdef.org
cdig.info
cdjp.org
cdn-images.mailchimp.com
cdn-telegram.org
cdn.arstechnica.net
cdn.assets.lfpcontent.com
cdn.helixstudios.net
cdn.jwplayer.com
cdn.printfriendly.com
cdn.seatguru.com
cdn.softlayer.net
cdn.statically.io
cdn1.lp.saboom.com
cdnews.com.tw
cdninstagram.com
cdp1989.org
cdp1998.org
cdp2006.org
cdpeu.org
cdpuk.co.uk
cdpweb.org
cdpwu.org
cdw.com
cecc.gov
cellulo.info
cenews.eu
centauro.com.br
centerforhumanreprod.com
centralnation.com
centurys.net
certificate-transparency.org
certificate.revocationcheck.com
cfhks.org.hk
cfos.de
cfr.org
cftfc.com
cgdepot.org
cgst.edu
change.org
changeip.name
changeip.net
changeip.org
changp.com
channelnewsasia.com
chanworld.org
chaoex.com
chaos.social
character.ai
chat.lmsys.org
chatgpt.com
chatnook.com
chaturbate.com
checkgfw.com
chengmingmag.com
chenguangcheng.com
chenpokong.com
chenpokongvip.com
chenshan20042005.wordpress.com
cherrysave.com
chhongbi.org
china-mmm.jp.net
china-mmm.net
china-review.com.ua
china-week.com
china.ucanews.com
china101.com
china18.org
china21.com
china21.org
china5000.us
chinaaffairs.org
chinaaid.net
chinaaid.org
chinaaid.us
chinachange.org
chinachannel.hk
chinademocrats.org
chinadialogue.net
chinadigitaltimes.net
chinaelections.org
chinafile.com
chinafreepress.org
chinagate.com
chinagfw.org
chinagonet.com
chinahorizon.org
chinahush.com
chinainperspective.com
chinalaborwatch.org
chinalawandpolicy.com
chinalawtranslate.com
chinamule.com
chinamz.org
chinanewscenter.com
chinapost.com.tw
chinapress.com.my
chinarightsia.org
chinasmile.net
chinasocialdemocraticparty.com
chinasoul.org
chinasucks.net
chinatopsex.com
chinatown.com.au
chinauncensored.tv
chinaview.wordpress.com
chinaway.org
chinaworker.info
chinayouth.org.hk
chinese-leaders.org
chinese-memorial.org
chinese.donga.com
chinese.engadget.com
chinese.irib.ir
chinese.soifind.com
chinesedaily.com
chinesedailynews.com
chinesedemocracy.com
chinesegay.org
chinesen.de
chinesenews.net.au
chinesepen.org
chineseradioseattle.com
chineseupress.com
chingcheong.com
chinman.net
chithu.org
chobit.cc
chrdnet.com
christianfreedom.org
christianstudy.com
christiantimes.org.hk
christusrex.org
chrlawyers.hk
chrome.com
chromecast.com
chromeexperiments.com
chromestatus.com
chromium.org
chuang-yen.org
chubold.com
chubun.com
churchinhongkong.org
chushigangdrug.ch
ci-en.jp
cici.com
ciciai.com
cienen.com
cineastentreff.de
cipfg.org
cirosantilli.com
citizencn.com
citizenlab.ca
citizenlab.org
citizensradio.org
city365.ca
city9x.com
citypopulation.de
citytalk.tw
civicparty.hk
civilhrfront.org
civiliangunner.com
civilmedia.tw
civitai.com
cixiaoya.club
ck101.com
clarionproject.org
classicalguitarblog.net
claude.ai
clb.org.hk
cldr.unicode.org
cleansite.biz
cleansite.info
cleansite.us
clearharmony.net
clearsurance.com
clearwisdom.net
clementine-player.org
clinica-tibet.ru
clipconverter.cc
clipfish.de
cloud.dify.ai
cloud.mail.ru
cloudflare-dns.com
cloudflare-ipfs.com
cloudfunctions.net
club1069.com
clubhouseapi.com
clyp.it
cmcn.org
cmegroup.com
cmi.org.tw
cmp.hku.hk
cms.gov
cmule.com
cmx.im
cn-proxy.com
cn.fmnnow.com
cn.freeones.com
cn.nytstyle.com
cn.sandscotaicentral.com
cn.shafaqna.com
cn.streetvoice.com
cn.theaustralian.com.au
cn.uncyclopedia.wikia.com
cn.uptodown.com
cn6.eu
cna.com.tw
cnabc.com
cnbbnews.wordpress.com
cnbeta.com.tw
cnd.org
cnex.org.cn
cnineu.com
cnn.com
cnnews.chosun.com
cnpolitics.org
cnproxy.com
co.ng.mil
coat.co.jp
cobinhood.com
cochina.org
codeshare.io
codeskulptor.org
cofacts.tw
coffeemanga.to
coinbase.com
coinbene.com
coinex.com
coingecko.com
coingi.com
coinmarketcap.com
coinrail.co.kr
cointiger.com
cointobe.com
coinut.com
colacloud.net
collateralmurder.com
collateralmurder.org
comefromchina.com
comic-mega.me
commandarms.com
comments.app
commentshk.com
communistcrimes.org
communitychoicecu.com
comparitech.com
compileheart.com
compress.to
connect.facebook.net
conoha.jp
contactmagazine.net
contests.twilio.com
convio.net
cool18.com
coolaler.com
coolder.com
coolloud.org.tw
coolncute.com
coolstuffinc.com
copilot.microsoft.com
corumcollege.com
cos-moe.com
cosplayjav.pl
costco.com
cotweet.com
counter.social
coursehero.com
covenantswatch.org.tw
coze.com
cpj.org
cpu-monkey.com
cq99.us
crackle.com
crazypool.org
crazys.cc
crazyshit.com
crbug.com
crchina.org
crd-net.org
creaders.net
creadersnet.com
creativelab5.com
cristyli.com
crocotube.com
crossfire.co.kr
crossvpn.net
crosswall.org
croxyproxy.com
crrev.com
crucial.com
crunchyroll.com
cruxpool.com
csdparty.com
csis.org
csmonitor.com
css.pixnet.in
csuchen.de
csw.org.uk
ct.org.tw
ctao.org
ctinews.com
ctitv.com.tw
ctowc.org
cts.com.tw
ctwant.com
cuhkacs.org
cuiweiping.net
culture.tw
cumlouder.com
curvefish.com
cusp.hk
cutout.pro
cutscenes.net
cw.com.tw
cyberghost.natado.com
cyberghostvpn.com
cynscribe.com
d-fukyu.com
d.cash
d100.net
d2bay.com
d2pass.com
dabr.co.uk
dabr.eu
dabr.me
dabr.mobi
dadazim.com
dadi360.com
dafabet.com
dafagood.com
dafahao.com
dafoh.org
daftporn.com
dagelijksestandaard.nl
daidostup.ru
dailymail.co.uk
dailymotion.com
dailynews.sina.com
dailysabah.com
dailyview.tw
dajiyuan.com
dajiyuan.de
dajiyuan.eu
dalailama-archives.org
dalailama.com
dalailama.mn
dalailama.ru
dalailama80.org
dalailamacenter.org
dalailamafellows.org
dalailamafilm.com
dalailamafoundation.org
dalailamahindi.com
dalailamainaustralia.org
dalailamajapanese.com
dalailamaprotesters.info
dalailamaquotes.org
dalailamatrust.org
dalailamavisit.org.nz
dalailamaworld.com
dalianmeng.org
daliulian.org
danbooru.donmai.us
danke4china.net
daodu14.jigsy.com
daolan.net
darktech.org
darpa.mil
darrenliuwei.com
dashlane.com
data-vocabulary.org
data.gov.tw
daum.net
david-kilgour.com
dawangidc.com
daxa.cn
daylife.com
db.tt
dcard.tw
dcmilitary.com
ddc.com.tw
ddex.io
ddns.info
ddns.me.uk
ddns.mobi
ddns.ms
ddns.name
ddns.net
ddns.us
deadhouse.org
deadline.com
deaftone.com
debug.com
deck.ly
deck.new
decodet.co
deepai.org
deepmind.com
deezer.com
definebabe.com
deja.com
delcamp.net
delicious.com
demo.unlock-music.dev
democrats.org
demosisto.hk
deno.dev
depositphotos.com
desc.se
desipro.de
dessci.com
destiny.xfiles.to
destroy-china.jp
deutsche-welle.de
developers.box.net
deviantart.com
deviantart.net
devio.us
devpn.com
devv.ai
dfn.org
dharamsalanet.com
dharmakara.net
diaoyuislands.org
difangwenge.org
digiland.tw
digisfera.com
diigo.com
dipity.com
directcreative.com
discoins.com
disconnect.me
discord.com
discord.gg
discordapp.com
discordapp.net
discuss.com.hk
discuss4u.com
dish.com
disk.yandex.com
disk.yandex.ru
disneyplus.com
disp.cc
disqus.com
dit-inc.us
diyin.org
dizhidizhi.com
dizhuzhishang.com
djangosnippets.org
dl-laby.jp
dl.box.net
dlive.tv
dlsite.com
dlyoutube.com
dm530.net
dma.mil
dmc.nico
dmcdn.net
dmhy.org
dmm.co.jp
dns-dns.com
dns-stuff.com
dns04.com
dns05.com
dns1.us
dns2.us
dns2go.com
dnscrypt.org
dnset.com
dnsrd.com
dnssec.net
dnvod.tv
doc.new
docker.com
docker.io
docs.deno.com
docs.new
doctorvoice.org
documentingreality.com
dogfartnetwork.com
dojin.com
dolc.de
dolf.org.hk
domain.club.tw
domaintoday.com.au
dongtaiwang.com
dongtaiwang.net
dongyangjing.com
dontfilter.us
doom9.org
doosho.com
doourbest.org
dorjeshugden.com
dotplane.com
dotsub.com
dotvpn.com
doub.io
doubibackup.com
doubiyunbackup.com
doublethinklab.org
douchi.space
dougscripts.com
doujincafe.com
download.aircrack-ng.org
download.cnet.com
dphk.org
dpool.top
dpp.org.tw
dpr.info
dragonex.io
dragonsprings.org
dreamamateurs.com
drepung.org
drgan.net
dropbooks.tv
dropbox.com
dropboxapi.com
dropboxusercontent.com
drtuber.com
dscn.info
dsmtp.com
dstk.dk
dtiblog.com
dtic.mil
dubox.com
duck.com
duckduckgo.com
duckload.com
duckmylife.com
duga.jp
duihua.org
duihuahrjournal.org
dumb1.com
duping.net
duplicati.com
dupola.com
dupola.net
dushi.ca
duyaoss.com
dvdpac.com
dvorak.org
dw-world.com
dw-world.de
dw.com
dw.de
dweb.link
dwnews.com
dwnews.net
dynamic-dns.net
dynamicdns.biz
dynamicdns.co.uk
dynamicdns.me.uk
dynamicdns.org.uk
dynawebinc.com
dyndns-ip.com
dyndns-pics.com
dyndns.org
dyndns.pro
dynssl.com
dynu.com
dynu.net
dynupdate.no-ip.com
dysfz.cc
dzze.com
e-classical.com.tw
e-gold.com
e-hentai.org
e-hentaidb.com
e-info.org.tw
e-zone.com.hk
e123.hk
e621.net
earlytibet.com
earthcam.com
earthvpn.com
eastasiaforum.org
easternlightning.org
eastturkestan.com
eastturkistan-gov.org
eastturkistancc.org
eastturkistangovernmentinexile.us
easyca.ca
easypic.com
ebony-beauty.com
ebook.hyread.com.tw
ebookbrowse.com
ebookee.com
ebtcbank.com
ecfa.org.tw
echofon.com
ecimg.tw
ecministry.net
economist.com
edgecastcdn.net
edicypages.com
edmontonchina.cn
edmontonservice.com
edoors.com
edubridge.com
edupro.org
edx-cdn.org
eeas.europa.eu
eesti.ee
eevpn.com
efcc.org.hk
effers.com
efksoft.com
efukt.com
eic-av.com
eireinikotaerukai.com
eisbb.com
eksisozluk.com
elconfidencial.com
electionsmeter.com
elgoog.im
elpais.com
eltondisney.com
emaga.com
emanna.com
emilylau.org.hk
empfil.com
emule-ed2k.com
emulefans.com
emuparadise.me
en.favotter.net
en.hao123.com
enanyang.my
encrypt.me
encyclopedia.com
enewstree.com
enfal.de
englishforeveryone.org
englishfromengland.co.uk
englishpen.org
enlighten.org.tw
entermap.com
epac.to
episcopalchurch.org
epochhk.com
epochtimes-bg.com
epochtimes-romania.com
epochtimes.co.il
epochtimes.co.kr
epochtimes.com
epochtimes.com.tw
epochtimes.cz
epochtimes.de
epochtimes.fr
epochtimes.it
epochtimes.jp
epochtimes.ru
epochtimes.se
epochtimestr.com
epochweek.com
epochweekly.com
eporner.com
equinenow.com
erabaru.net
eracom.com.tw
eraysoft.com.tr
erepublik.com
erights.net
eriversoft.com
ernestmandel.org
erodaizensyu.com
erodoujinlog.com
erodoujinworld.com
eromanga-kingdom.com
eromangadouzin.com
eromon.net
eroprofile.com
eroticsaloon.net
esg.t91y.com
eslite.com
esmtp.biz
esurance.com
etaa.org.au
etadult.com
etaiwannews.com
etherdelta.com
ethermine.org
etherscan.com
etherscan.io
etizer.org
etokki.com
etools.ncol.com
etowns.net
etowns.org
etsy.com
ettoday.net
etvonline.hk
eucasino.com
eulam.com
eurekavpt.com
euronews.com
evchk.wikia.com
everipedia.org
evschool.net
exam.gov.tw
exblog.jp
exchristian.hk
exhentai.org
exmo.com
exmormon.org
expatshield.com
expecthim.com
expekt.com
experts-univers.com
exploader.net
expressvpn.com
exrates.me
extmatrix.com
extremetube.com
ey.gov.tw
eyevio.jp
eyny.com
ezpeer.com
ezua.com
f-droid.org
f2pool.com
f8.com
fa.gov.tw
facebook.com
facebook.de
facebook.design
facebook.hu
facebook.in
facebook.nl
facebook.se
facebookmail.com
facebookquotes4u.com
faceless.me
facesofnyfw.com
facesoftibetanselfimmolators.info
factchecklab.org
factpedia.org
faith100.org
faithfuleye.com
faiththedog.info
fakku.net
fallenark.com
falsefire.com
falun-co.org
falun-ny.net
falunart.org
falunasia.info
falunau.org
falunaz.net
falundafa-dc.org
falundafa-florida.org
falundafa-nc.org
falundafa-pa.net
falundafa.org
falundafaindia.org
falundafamuseum.org
falungong.club
falungong.de
falungong.org.uk
falunhr.org
faluninfo.de
faluninfo.net
falunpilipinas.net
familyfed.org
famunion.com
fan-qiang.com
fanbox.cc
fangeming.com
fangeqiang.com
fanglizhi.info
fangong.forums-free.com
fangong.org
fangongheike.com
fanhaodang.com
fanhaolou.com
fanqiang.network
fanqiang.tk
fanqiangdang.com
fanqianghou.com
fanqiangzhe.com
fanswong.com
fantv.hk
fanyue.info
fapdu.com
faproxy.com
faqserv.com
fartit.com
farwestchina.com
fast.com
fast.wistia.com
fastestvpn.com
fastpic.ru
fastssh.com
faststone.org
fatbtc.com
favstar.fm
fawanghuihui.org
faydao.com
faz.net
fb.com
fb.me
fb.watch
fbaddins.com
fbcdn.net
fbsbx.com
fbworkmail.com
fc2.com
fc2blog.net
fc2china.com
fc2cn.com
fda.gov.tw
fdc64.de
fdc64.org
fdc89.jp
feedburner.com
feeder.co
feedly.com
feeds.fileforum.com
feedx.net
feelssh.com
feer.com
feitian-california.org
feitianacademy.org
feixiaohao.com
feministteacher.com
fengzhenghu.com
fengzhenghu.net
fevernet.com
ff.im
fffff.at
fflick.com
ffvpn.com
fgmtv.net
fgmtv.org
fhreports.net
fiddle.jshell.net
figprayer.com
fileflyer.com
fileserve.com
filesor.com
fillthesquare.org
filmingfortibet.org
filthdump.com
financetwitter.com
financialexpress.com
finchvpn.com
findmespot.com
findyoutube.com
findyoutube.net
fingerdaily.com
firearmsworld.net
firebaseio.com
fireofliberty.info
fireofliberty.org
firetweet.io
firstpost.com
firstrade.com
fish.audio
flagsonline.it
flecheinthepeche.fr
fleshbot.com
fleursdeslettres.com
flexpool.io
flgjustice.org
flickr.com
flickrhivemind.net
flickriver.com
fling.com
flipboard.com
flipkart.com
flitto.com
flnet.org
flog.tw
flowhongkong.net
flypool.org
flyvpn.com
flyzy2005.com
fnac.be
fnac.com
fnc.ebc.net.tw
fochk.org
focustaiwan.tw
focusvpn.com
fofg.org
fooooo.com
forbes.com
foreignaffairs.com
foreignpolicy.com
form.new
forms.new
forum.baby-kingdom.com
forum.cyberctm.com
forum.mymaji.com
forum.palmislife.com
forum.slime.com.tw
forum.tvb.com
forum.xinbao.de
forum4hk.com
fountmedia.io
fourthinternational.org
foxgay.com
foxsub.com
foxtang.com
fpmt-osel.org
fpmt.org
fpmt.tw
fpmtmexico.org
fq.wikia.com
fqrouter.com
frank2019.me
franklc.com
freakshare.com
free-gate.org
free-hada-now.org
free-proxy.cz
free-ss.site
free.bg
free.com.tw
free.fr
freebeacon.com
freebrowser.org
freechal.com
freechina.net
freechina.news
freeddns.com
freeddns.org
freedomcollection.org
freedomhouse.org
freedominfonetweb.wordpress.com
freedomsherald.org
freeforums.org
freegao.com
freehongkong.org
freeilhamtohti.org
freekazakhs.org
freelotto.com
freeman2.com
freemoren.com
freemorenews.com
freemuse.org
freenet-china.org
freenetproject.org
freenewscn.com
freeopenvpn.com
freeoz.org
freess.org
freessh.us
freetcp.com
freetibet.net
freetibet.org
freetibetanheroes.org
freetls.fastly.net
freetribe.me
freeviewmovies.com
freevpn.me
freevpn.nl
freewallpaper4.me
freewebs.com
freewechat.com
freewww.info
freexinwen.com
freeyellow.com
freezhihu.org
friendfeed.com
friends-of-tibet.org
friendsoftibet.org
fril.jp
fring.com
fringenetwork.com
from-pr.com
from-sd.com
fromchinatousa.net
frommel.net
frontlinedefenders.org
frootvpn.com
froth.zone
fscked.org
fsurf.com
ft.com
ftchinese.com
ftp1.biz
ftpserver.biz
ftv.com.tw
ftvnews.com.tw
ftx.com
fucd.com
fuchsia.dev
fuckccp.com
fuckccp.xyz
fuckgfw.org
fulione.com
fullerconsideration.com
fullservicegame.com
funf.tw
funkyimg.com
funp.com
fuq.com
furbo.org
furhhdl.org
furinkan.com
furrybar.com
futurechinaforum.org
futuremessage.org
fux.com
fuyindiantai.org
fuyu.org.tw
fw.cm
fxcm-chinese.com
fxnetworks.com
g-area.org
g-queen.com
g.co
g0v.social
g6hentai.com
gab.com
gabocorp.com
gaeproxy.com
gaforum.org
gagaoolala.com
galaxymacau.com
galenwu.com
galstars.net
game735.com
gamebase.com.tw
gamejolt.com
gamer.com.tw
gamez.com.tw
gamousa.com
ganges.com
ganjing.com
ganjingworld.com
gaoming.net
gaopi.net
gardennetworks.com
gardennetworks.org
gartlive.com
garudalinux.org
gate.io
gatecoin.com
gather.com
gatherproxy.com
gaybubble.com
gaycn.net
gayhub.com
gaymap.cc
gaymenring.com
gaytube.com
gaywatch.com
gazotube.com
gcc.org.hk
gclubs.com
gcmasia.com
gcpnews.com
gcr.io
gdaily.org
gdzf.org
geek-art.net
geekerhome.com
gekikame.com
gelbooru.com
generated.photos
genius.com
geocities.co.jp
geocities.com
geocities.jp
geph.io
gerefoundation.org
get.app
get.dev
get.how
get.page
getastrill.com
getchu.com
getcloak.com
getfoxyproxy.org
getgom.com
geti2p.net
getiton.com
getlantern.org
getmalus.com
getmdl.io
getoutline.org
getsession.org
getsocialscope.com
getsync.com
gettr.com
gettrials.com
getuploader.com
gfbv.de
gfsale.com
gfw.press
gfw.report
gfwatch.org
ggpht.com
ggssl.com
ghidra-sre.org
ghostpath.com
ghut.org
giantessnight.com
gifree.com
giga-web.jp
gigporno.ru
girlbanker.com
git.io
gitbook.io
github.blog
github.com
github.io
githubassets.com
githubcopilot.com
githubusercontent.com
gitlab.net
gizlen.net
gjczz.com
glarity.app
glass8.eu
global.ssl.fastly.net
globaljihad.net
globalmediaoutreach.com
globalmuseumoncommunism.org
globalrescue.net
globaltm.org
globalvoices.org
globalvoicesonline.org
globalvpn.net
glock.com
gloryhole.com
glorystar.me
gluckman.com
glype.com
gmail.com
gmgard.com
gmll.org
gmodules.com
gmp4.com
gnci.org.hk
gnews.org
go-to-zlibrary.se
go141.com
go5.dev
goagent.biz
godaddy.com
godfootsteps.org
godoc.org
godsdirectcontact.co.uk
godsdirectcontact.org
godsdirectcontact.org.tw
godsimmediatecontact.com
gofundme.com
gohappy.com.tw
gojet.krtco.com.tw
gokbayrak.com
golang.org
goldbet.com
goldbetsports.com
golden-ages.org
goldeneyevault.com
goldenfrog.com
goldstep.net
goldwave.com
gongm.in
goo.gl
goo.gle
goo.ne.jp
good.news
gooday.xyz
goodhope.school
goodnewsnetwork.org
goodreaders.com
goodreads.com
goodtv.com.tw
goodtv.tv
goofind.com
google.ad
google.ae
google.al
google.am
google.as
google.at
google.az
google.ba
google.be
google.bf
google.bg
google.bi
google.bj
google.bs
google.bt
google.by
google.ca
google.cat
google.cd
google.cf
google.cg
google.ch
google.ci
google.cl
google.cm
google.cn
google.co.ao
google.co.bw
google.co.ck
google.co.cr
google.co.id
google.co.il
google.co.in
google.co.jp
google.co.ke
google.co.kr
google.co.ls
google.co.ma
google.co.mz
google.co.nz
google.co.th
google.co.tz
google.co.ug
google.co.uk
google.co.uz
google.co.ve
google.co.vi
google.co.za
google.co.zm
google.co.zw
google.com
google.com.af
google.com.ag
google.com.ai
google.com.ar
google.com.au
google.com.bd
google.com.bh
google.com.bn
google.com.bo
google.com.br
google.com.bz
google.com.co
google.com.cu
google.com.cy
google.com.do
google.com.ec
google.com.eg
google.com.et
google.com.fj
google.com.gh
google.com.gi
google.com.gt
google.com.hk
google.com.jm
google.com.kh
google.com.kw
google.com.lb
google.com.ly
google.com.mm
google.com.mt
google.com.mx
google.com.my
google.com.na
google.com.nf
google.com.ng
google.com.ni
google.com.np
google.com.om
google.com.pa
google.com.pe
google.com.pg
google.com.ph
google.com.pk
google.com.pr
google.com.py
google.com.qa
google.com.sa
google.com.sb
google.com.sg
google.com.sl
google.com.sv
google.com.tj
google.com.tr
google.com.tw
google.com.ua
google.com.uy
google.com.vc
google.com.vn
google.cv
google.cz
google.de
google.dev
google.dj
google.dk
google.dm
google.dz
google.ee
google.es
google.fi
google.fm
google.fr
google.ga
google.ge
google.gg
google.gl
google.gm
google.gp
google.gr
google.gy
google.hn
google.hr
google.ht
google.hu
google.ie
google.im
google.iq
google.is
google.it
google.je
google.jo
google.kg
google.ki
google.kz
google.la
google.li
google.lk
google.lt
google.lu
google.lv
google.md
google.me
google.mg
google.mk
google.ml
google.mn
google.ms
google.mu
google.mv
google.mw
google.ne
google.nl
google.no
google.nr
google.nu
google.pl
google.pn
google.ps
google.pt
google.ro
google.rs
google.ru
google.rw
google.sc
google.se
google.sh
google.si
google.sk
google.sm
google.sn
google.so
google.sr
google.st
google.td
google.tg
google.tk
google.tl
google.tm
google.tn
google.to
google.tt
google.vg
google.vu
google.ws
googleapis.com
googleapps.com
googlearth.com
googleartproject.com
googleblog.com
googlebot.com
googlechinawebmaster.com
googlecode.com
googlecommerce.com
googledomains.com
googledrive.com
googleearth.com
googlefiber.net
googlegroups.com
googlehosted.com
googleideas.com
googleinsidesearch.com
googlemail.com
googlemashups.com
googlepagecreator.com
googleplay.com
googleplus.com
googlescholar.com
googlesource.com
googleusercontent.com
googlevideo.com
googleweblight.com
googlezip.net
gopetition.com
goreforum.com
goregrish.com
gospelherald.com
got-game.org
gotdns.ch
gotgeeks.com
gotquestions.org
gotrusted.com
gotw.ca
gov.ir
gov.taipei
gov.tw
gr8domain.biz
gr8name.biz
grammaly.com
grandtrial.org
graph.org
graphis.ne.jp
graphql.org
gravatar.com
greasyfork.org
greatfire.org
greatfire.us7.list-manage.com
greatfirewall.biz
greatfirewallofchina.org
greatroc.org
greatzhonghua.org
greenparty.org.tw
greenpeace.com.tw
greenpeace.org
greenreadings.com
greenvpn.net
greenvpn.org
grindr.com
grok.com
ground.news
gs-discuss.com
gsearch.media
gsp.target.com
gstatic.com
gtricks.com
gtv.org
gtv1.org
gu-chu-sum.org
guaguass.com
guancha.org
guangming.com.my
guardster.com
guishan.org
gumroad.com
gun-world.net
gunsamerica.com
gunsandammo.com
guruonline.hk
gutteruncensored.com
gvlib.com
gvm.com.tw
gvt1.com
gvt3.com
gwins.org
gwtproject.org
gyalwarinpoche.com
gyatsostudio.com
gzm.tv
gzone-anime.info
h-china.org
h-moe.com
h1n1china.org
h528.com
h5dm.com
h5galgame.me
hacken.cc
hacker.org
hackmd.io
hackthatphone.net
hahlo.com
haijiao.com
haiwaikan.com
hakkatv.org.tw
halktv.com.tr
handcraftedsoftware.org
hanime.tv
hanime1.me
hao.news
haproxy.org
hardsextube.com
hautelook.com
hautelookcdn.com
have8.com
hbg.com
hbo.com
hclips.com
hd.stheadline.com
hdtvb.net
hdzog.com
heartyit.com
heavy-r.com
hec.su
hecaitou.net
hechaji.com
heeact.edu.tw
hegre-art.com
helloandroid.com
helloqueer.com
hellouk.org
helpeachpeople.com
helpster.de
helpzhuling.org
hentai.to
hentaitube.tv
hentaivideoworld.com
heqinglian.net
heritage.org
herokuapp.com
herominers.com
hexieshe.com
hexieshe.xyz
hexxeh.net
heyuedi.com
heyzo.com
hgseav.com
hhdcb3office.org
hhthesakyatrizin.org
hi-on.org.tw
hiccears.com
hidden-advent.org
hide.me
hidecloud.com
hideipvpn.com
hideman.net
hideme.nl
hidemy.name
hidemyass.com
hidemycomp.com
higfw.com
highpeakspureearth.com
highrockmedia.com
hiitch.com
hikinggfw.org
hilive.tv
himalayan-foundation.org
himalayanglacier.com
himemix.com
hindustantimes.com
hinet.net
hitbtc.com
hitomi.la
hiveon.net
hiwifi.com
hizb-ut-tahrir.info
hizb-ut-tahrir.org
hizbuttahrir.org
hjclub.info
hk-pub.com
hk.frienddy.com
hk.geocities.com
hk.gradconnection.com
hk.hao123img.com
hk.jiepang.com
hk01.com
hka8964.wordpress.com
hkacg.com
hkacg.net
hkanews.wordpress.com
hkatvnews.com
hkbc.net
hkbf.org
hkbookcity.com
hkchronicles.com
hkchurch.org
hkci.org.hk
hkcmi.edu
hkcnews.com
hkcoc.com
hkcoc.weather.com.hk
hkdailynews.com.hk
hkday.net
hkdc.us
hkdf.org
hkej.com
hkepc.com
hket.com
hkfaa.com
hkfront.org
hkgalden.com
hkgolden.com
hkgpao.com
hkheadline.com
hkhkhk.com
hkhrc.org.hk
hkjc.com
hkjp.org
hklft.com
hklts.org.hk
hkmap.live
hkopentv.com
hkpeanut.com
hkptu.org
hkreporter.com
hkreporter.loved.hk
hmoegirl.com
hmv.co.jp
hmvdigital.ca
hmvdigital.com
hnjhj.com
hnntube.com
hojemacau.com.mo
hola.com
hola.org
hole.thu.monster
holyspiritspeaks.org
home.saxo
home.sina.com
homedepot.com
homeperversion.com
homeservershow.com
hongkong.fandom.com
hongkongfp.com
hongmeimei.com
hongzhi.li
honven.xyz
hootsuite.com
hoover.org
hoovers.com
hopto.org
hornygamer.com
hornytrip.com
horrorporn.com
hostloc.com
hotair.com
hotav.tv
hotcoin.com
hotels.cn
hotfrog.com.tw
hotgoo.com
hotpot.hk
hotshame.com
hotspotshield.com
hottg.com
hotvpn.com
howtoforge.com
hoxx.com
hoy.tv
hpa.gov.tw
hpjav.com
hqcdp.org
hqjapanesesex.com
hqmovies.com
hqsbnet.wordpress.com
hqsbonline.wordpress.com
hrcchina.org
hrea.org
hrichina.org
hrntt.org
hrtsea.com
hrw.org
hrweb.org
hsex.men
hsjp.net
hsselite.com
hst.net.tw
hstern.net
hstt.net
ht.ly
htkou.net
htl.li
html5rocks.com
https443.net
https443.org
huaglad.com
huanghuagang.org
huangyiyu.com
huaren.us
huaren4us.com
huashangnews.com
huaxiabao.org
huaxin.ph
huayuworld.org
hudatoriq.web.id
hudson.org
huffingtonpost.com
huffpost.com
huggingface.co
hugoroy.eu
huhaitai.com
huhamhire.com
huhangfei.com
hulkshare.com
hulu.com
huluim.com
humanparty.me
humanrightspressawards.org
hung-ya.com
huobi.co
huobi.com
huobi.me
huobi.pro
huobi.sc
huobipool.com
huobipro.com
huping.net
hurgokbayrak.com
hurriyet.com.tr
hustlercash.com
hut2.ru
hutianyi.net
hutong9.net
huyandex.com
hwadzan.tw
hwayue.org.tw
hxwk.org
hxwq.org
hybrid-analysis.com
hyperrate.com
hypothes.is
i-cable.com
i-part.com.tw
i-scmp.com
i.111666.best
i.lithium.com
i2p2.de
i818hk.com
iam.soy
iamtopone.com
iask.ca
iav19.com
iavian.net
ibiblio.org
ibros.org
ibvpn.com
icams.com
ice.audionow.com
icedrive.net
icij.org
icl-fi.org
icoco.com
iconpaper.org
icu-project.org
id.hao123.com
id.heroku.com
idemocracy.asia
identi.ca
idiomconnection.com
idope.se
idouga.com
idv.tw
ied2k.net
ienergy1.com
ifan.cz.cc
ifcss.org
ifjc.org
ifreechina.wordpress.com
ifreewares.com
ift.tt
igcd.net
igfw.net
igmg.de
igoogle.com
igotmail.com.tw
igvita.com
ihao.org
iicns.com
iipdigital.usembassy.gov
ikstar.com
ikwb.com
ilbe.com
ilhamtohtiinstitute.org
illawarramercury.com.au
illusionfactory.com
ilove80.be
ilovelongtoes.com
im88.tw
imageab.com
imagefap.com
imageflea.com
imageglass.org
images-gaytube.com
imageshack.us
imagevenue.com
imagezilla.net
imago-images.com
imb.org
imdb.com
img.dlsite.jp
img.ly
img.picgo.net
imgasd.com
imgchili.net
imgmega.com
imgur.com
imkev.com
imlive.co
imlive.com
immigration.gov.tw
immoral.jp
impact.org.au
improd.works
in-disguise.com
in99.org
incapdns.net
incloak.com
incredibox.fr
independent.co.uk
india.com
indiablooms.com
indianarrative.com
indiandefensenews.in
indiatoday.in
indiemerch.com
inews-api.tvb.com
info-graf.fr
infura.io
inherit.live
initiativesforchina.org
inkbunny.net
inkui.com
inmediahk.net
inoreader.com
inote.tw
insecam.org
inside.com.tw
insidevoa.com
instagram.com
instanthq.com
institut-tibetain.org
interactivebrokers.com
internet.org
internetdefenseleague.org
internetfreedom.org
internetpopculture.com
inthenameofconfuciusmovie.com
investigating.wordpress.com
invidio.us
inxian.com
iownyour.org
ipdefenseforum.com
ipfire.org
ipfs.4everland.io
ipfs.io
iphone4hongkong.com
iphonetaiwan.org
iphonix.fr
ipicture.ru
ipify.org
ipjetable.net
ipobar.com
ipoock.com
iportal.me
ippotv.com
ipredator.se
iptv.com.tw
iptvbin.com
ipvanish.com
irangov.ir
iredmail.org
irna.ir
ironpython.net
ironsocket.com
is-a-hunter.com
is.gd
isaacmao.com
isasecret.com
isc.sans.edu
isgreat.org
ishr.ch
islahhaber.net
islam.org.hk
islamawareness.net
islamhouse.com
islamicity.com
islamicpluralism.org
islamtoday.net
ismaelan.com
ismalltits.com
ismprofessional.net
isohunt.com
israbox.com
issuu.com
istars.co.nz
istockphoto.com
isunaffairs.com
isuntv.com
isupportuyghurs.org
itaiwan.gov.tw
italiatibet.org
itasoftware.com
itch.io
itemdb.com
itemfix.com
ithelp.ithome.com.tw
itiger.com
itsaol.com
itshidden.com
itsky.it
itweet.net
iu45.com
iuhrdf.org
iuksky.com
ivacy.com
ivonblog.com
ivpn.net
iwara.tv
ixquick.com
ixxx.com
iyouport.com
iyouport.org
izaobao.us
izles.net
izlesem.org
j.mp
jable.tv
jamaat.org
jamestown.org
jamyangnorbu.com
jan.ai
japan-whores.com
japanhdv.com
japantimes.co.jp
jav.com
jav101.com
jav321.com
jav68.tv
javakiba.org
javbus.com
javdb.com
javfinder.ai
javfor.me
javhd.com
javhip.com
javhub.net
javhuge.com
javlibrary.com
javmobile.net
javmoo.com
javmoo.xyz
javseen.com
javtag.com
javzoo.com
jbtalks.cc
jbtalks.com
jbtalks.my
jcpenney.com
jdwsy.com
jeanyim.com
jetos.com
jex.com
jgoodies.com
jiangweiping.com
jiaoyou8.com
jichangtj.com
jiehua.cz
jieshibaobao.com
jifangge.com
jigglegifs.com
jihadintel.meforum.org
jihadology.net
jiji.com
jims.net
jingpin.org
jingsim.org
jinpianwang.com
jinrizhiyi.news
jintian.net
jinx.com
jitouch.com
jjgirls.com
jkb.cc
jkforum.net
jkub.com
jma.go.jp
jmsc.hku.hk
jmscult.com
joachims.org
jobnewera.wordpress.com
joinclubhouse.com
joinmastodon.org
jornaldacidadeonline.com.br
journalchretien.net
journalofdemocracy.org
joymiihub.com
joyourself.com
jp.hao123.com
jpl.nasa.gov
jpopforum.net
jsdelivr.net
jtvnw.net
judicial.gov.tw
juhuaren.com
jukujo-club.com
juliepost.com
juliereyc.com
junauza.com
june4commemoration.org
jungleheart.com
juoaa.com
justdied.com
justfreevpn.com
justhost.ru
justmysocks.net
justmysocks1.net
justmysockscn.com
justpaste.it
justtristan.com
juziyue.com
jwmusic.org
jyxf.net
ka-wai.com
kadokawa.co.jp
kagyu.org
kagyu.org.za
kagyumonlam.org
kagyunews.com.hk
kagyuoffice.org
kagyuoffice.org.tw
kaiyuan.de
kakao.com
kanald.com.tr
kankan.today
kannewyork.com
kanshifang.com
kantie.org
kanzhongguo.com
kanzhongguo.eu
kaotic.com
karayou.com
karmapa-teachings.org
karmapa.org
kawaiikawaii.jp
kawase.com
kb.monitorware.com
kba-tx.org
kcoolonline.com
kebrum.com
kechara.com
keepandshare.com
keezmovies.com
kendatire.com
kenengba.com
kepard.com
kex.com
keycdn.com
khabdha.org
khatrimaza.org
kichiku-doujinko.com
kik.com
killwall.com
kindle4rss.com
kindleren.com
kineox.free.fr
kingdomsalvation.org
kinghost.com
kingkong.com.tw
kingstone.com.tw
kink.com
kinmen.org.tw
kinmen.travel
kinokuniya.com
kir.jp
kissbbao.cn
kiwi.kz
kk-whys.co.jp
kkbox.com
kknews.cc
kmuh.org.tw
knowledgerush.com
knowyourmeme.com
kobo.com
kobobooks.com
kodingen.com
kompozer.net
konachan.com
kone.com
koolsolutions.com
koornk.com
koranmandarin.com
korenan2.com
kpkuang.org
kqes.net
kraken.com
ksdl.org
ksnews.com.tw
kspcoin.com
ktzhk.com
kuaichedao.co
kucoin.com
kui.name
kukuku.uk
kun.im
kurashsultan.com
kurtmunger.com
kusocity.com
kwcg.ca
kwongwah.com.my
kxsw.life
kyofun.com
kyohk.net
kzaobao.com
kzeng.info
la-forum.org
lab.skk.moe
labiennale.org
ladbrokes.com
lagranepoca.com
lala.im
lalulalu.com
lama.com.tw
lamayeshe.com
lamnia.co.uk
lamrim.com
landofhope.tv
lanterncn.cn
lantosfoundation.org
laod.cn
laogai.org
laogairesearch.org
laomiu.com
laoyang.info
laqingdan.net
larsgeorge.com
lastcombat.com
lastfm.es
latelinenews.com
lausan.hk
lbank.info
ld.hao123img.com
ldplayer.net
ldplayer.tw
le-vpn.com
leafyvpn.net
lecloud.net
ledger.com
leeao.com.cn
lefora.com
left21.hk
legalporno.com
legra.ph
legsjapan.com
leisurecafe.ca
leisurepro.com
lematin.ch
lemonde.fr
lenwhite.com
lesoir.be
letou.com
letscorp.net
lflink.com
lflinkup.com
lflinkup.net
lflinkup.org
lhakar.org
lhasocialwork.org
li.taipei
liangyou.net
liangzhichuanmei.com
lianyue.net
liaowangxizang.net
liberal.org.hk
libertysculpturepark.com
libertytimes.com.tw
library.usc.cuhk.edu.hk
libredd.it
lifemiles.com
lighten.org.tw
lighti.me
lightnovel.cn
lightyearvpn.com
lih.kg
lihkg.com
like.com
lilaoshibushinilaoshi.com
limiao.net
line-apps.com
line-scdn.net
line.me
line.naver.jp
linear-abematv.akamaized.net
linglingfa.com
lingualeo.com
lingvodics.com
link-o-rama.com
linkedin.com
linkideo.com
linktr.ee
linkuswell.com
linux.org.hk
linuxtoy.org
lionsroar.com
lipuman.com
liquiditytp.com
liquidvpn.com
listennotes.com
listentoyoutube.com
listorious.com
lists.w3.org
litenews.hk
liu-xiaobo.org
liuhanyu.com
liuxiaobo.net
liuxiaotong.com
livecoin.net
livedoor.jp
liveleak.com
livemint.com
livestream.com
livevideo.com
livingstream.com
lizhizhuangbi.com
lkcn.net
load.to
lobsangwangyal.com
localbitcoins.com
localdomain.ws
localpresshk.com
lockestek.com
login.target.com
logos.com.hk
londonchinese.ca
longhair.hk
longmusic.com
longtermly.net
longtoes.com
lookpic.com
looktoronto.com
lotsawahouse.org
lotuslight.org.tw
lovetvshow.com
lpsg.com
lrfz.com
lrip.org
lsd.org.hk
lsforum.net
lsm.org
lsmchinese.org
lsmkorean.org
lsmradio.com
lsmwebcast.com
lsxszzg.com
ltn.com.tw
luckydesigner.space
luckymobile.ca
ludepress.com
luke54.com
luke54.org
lupm.org
lushstories.com
luxebc.com
lvhai.org
lvv2.com
ly.gov.tw
lyfhk.net
lzjscript.com
lzmtnews.org
m-team.cc
m.hkgalden.com
m.me
m.moegirl.org
ma.hao123.com
macgamestore.com
macrovpn.com
mad-ar.ch
madewithcode.com
madonna-av.com
madou.club
madrau.com
madthumbs.com
mahabodhi.org
mahjongsoul.com
maiio.net
mail-archive.com
main-ecnpaper-economist.content.pugpig.com
maiplus.com
maizhong.org
makemymood.com
makkahnewspaper.com
malaysiakini.com
mamingzhe.com
manchukuo.net
mandiant.com
mangabz.com
mangafox.com
mangafox.me
mangmang.run
maniash.com
manicur4ik.ru
mansion.com
mansionpoker.com
manta.com
manyvoices.news
maplew.com
marc.info
marguerite.su
martau.com
martsangkagyuofficial.org
maruta.be
marxist.com
marxist.net
marxists.org
mash.to
maskedip.com
mastodon.cloud
mastodon.host
mastodon.online
mastodon.social
mastodon.xyz
matainja.com
material.io
matome-plus.com
matome-plus.net
matrix.org
matters.news
matters.town
matthewdgreen.wordpress.com
mattwilcox.net
maven.neoforged.net
maxing.jp
mayimayi.com
mcadforums.com
mcaf.ee
mcfog.com
mcreasite.com
mcusercontent.com
md-t.org
me.me
me.ns.ci
me.youthwant.com.tw
meansys.com
media.nu.nl
media.org.hk
mediachinese.com
mediafire.com
mediafreakcity.com
mediawiki.org
medium.com
meetav.com
meetup.com
mefeedia.com
mefound.com
mega.co.nz
mega.io
mega.nz
megalodon.jp
megaproxy.com
megurineluka.com
meizhong.blog
meizhong.report
meltoday.com
memehk.com
memes.tw
memri.org
memrijttm.org
mercari.com
mercari.jp
mercatox.com
mercdn.net
mercyprophet.org
mergersandinquisitions.com
meridian-trust.org
meripet.com
merit-times.com.tw
mesotw.com
messenger.com
meta.com
metafilter.com
metamask.io
metart.com
metarthunter.com
meteorshowersonline.com
metro.taipei
metrohk.com.hk
metrolife.ca
metroradio.com.hk
mewe.com
mfxmedia.com
mgoon.com
mgstage.com
mh4u.org
mhradio.org
microvpn.com
middle-way.net
mihr.com
mihua.org
mikanani.me
mike.cz.cc
mikesoltys.com
milph.net
milsurps.com
mimiai.net
mimivip.com
mindrolling.org
mingdemedia.org
minghui-school.org
minghui.or.kr
minghui.org
minghuiyw.wordpress.com
mingjinglishi.com
mingjingnews.com
mingjingtimes.com
mingpao.com
mingpaocanada.com
mingpaomonthly.com
mingpaonews.com
mingpaony.com
mingpaosf.com
mingpaotor.com
mingpaovan.com
mingshengbao.com
minhhue.net
miningpoolhub.com
ministrybooks.org
minzhuzhongguo.org
miraheze.org
miroguide.com
mirror.xyz
mirrorbooks.com
mirrormedia.mg
missav.com
missav.ws
mist.vip
mitbbs.com
miuipolska.pl
mixero.com
mixi.jp
mixpod.com
mixx.com
mizzmona.com
mjib.gov.tw
mjlsh.usc.cuhk.edu.hk
mlc.ai
mlcool.com
mlzs.work
mm-cg.com
mmaaxx.com
mmmca.com
mnewstv.com
mo.nightlife141.com
mobatek.net
mobile01.com
mobileways.de
moby.to
mobypicture.com
mod.io
modernchinastudies.org
moeaic.gov.tw
moeerolibrary.com
moeshare.cc
mofa.gov.tw
mofos.com
mog.com
mohu.club
mohu.rocks
moj.gov.tw
mojim.com
mol.gov.tw
molihua.org
momoshop.com.tw
mondex.org
money-link.com.tw
moneydj.com
moneyhome.biz
monica.im
monitorchina.org
monocloud.me
monster.com
moodyz.com
moon.fm
moonbbs.com
moonbingo.com
moptt.tw
moresci.sale
morningsun.org
mos.ru
motherless.com
motor4ik.ru
mousebreaker.com
movements.org
moviefap.com
mp3buscador.com
mpettis.com
mpfinance.com
mpinews.com
mponline.hk
mrbasic.com
mrbonus.com
mrface.com
mrslove.com
mrtweet.com
msa-it.org
msguancha.com
msha.gov
mstdn.social
mswe1.org
mthruf.com
mubi.com
muchosucko.com
mullvad.net
multiply.com
multiproxy.org
multiupload.com
mummysgold.com
musicade.net
musixmatch.com
muslimvideo.com
muzi.com
muzi.net
muzu.tv
mvdis.gov.tw
mvg.jp
mx.hao123.com
mx981.com
my-formosa.com
my-private-network.co.uk
my-proxy.com
my.mail.ru
my.opera.com
my.pcloud.com
my03.com
myactimes.com
myaudiocast.com
myav.com.tw
mybbs.us
mybet.com
myca168.com
mycanadanow.com
mychinanet.com
mychinanews.com
mychinese.news
mycnnews.com
mycould.com
mydad.info
myddns.com
myeasytv.com
myeclipseide.com
myfreecams.com
myfreepaysite.com
myfreshnet.com
myftp.info
myiphide.com
myjs.tw
mykomica.org
mylftv.com
mymoe.moe
mymom.info
mynetav.net
mynetav.org
mynumber.org
myparagliding.com
mypicture.info
mypikpak.com
mypop3.net
mypop3.org
mypopescu.com
myreadingmanga.info
mysecondarydns.com
myshare.url.com.tw
mysinablog.com
mysite.verizon.net
myspace.com
myspacecdn.com
mytalkbox.com
mytizi.com
mywife.cc
mywww.biz
myz.info
naacoalition.org
naitik.net
nakedsecurity.sophos.com
nakido.com
nakuz.com
nalandabodhi.org
nalandawest.org
namgyal.org
namgyalmonastery.org
nanopool.org
nanyang.com
nanyangpost.com
nanzao.com
naol.ca
naol.cc
nat.gov.tw
nat.moe
national-lottery.co.uk
nationalawakening.org
nationalinterest.org
nationalreview.com
nationsonline.org
nationwide.com
naughtyamerica.com
naver.com
navyfamily.navy.mil
navyreserve.navy.mil
naweeklytimes.com
nbcnews.com
nbtvpn.com
nbyy.tv
nccwatch.org.tw
nch.com.tw
nchrd.org
ncn.org
nde.de
ndi.org
ndr.de
ned.org
nekoslovakia.net
nemesis2.qx.net
neo-miracle.com
neowin.net
netalert.me
netbirds.com
netcolony.com
netflav.com
netflix.com
netflix.net
netlify.app
netme.cc
netsarang.com
netsneak.com
network54.com
networkedblogs.com
networktunnel.net
new-3lunch.net
new-akiba.com
new96.ca
newcenturymc.com
newcenturynews.com
newchen.com
newgrounds.com
newhighlandvision.com
newindianexpress.com
newipnow.com
newlandmagazine.com.au
newmitbbs.com
newnews.ca
news.cnyes.com
news.ebc.net.tw
news.msn.com.tw
news.mt.co.kr
news.nationalgeographic.com
news.seehua.com
news.sina.com.hk
news.sinchew.com.my
news.singtao.ca
news.tvbs.com.tw
news.ycombinator.com
news1.kr
news100.com.tw
news18.com
newsancai.com
newsblur.com
newschinacomment.org
newsdetox.ca
newsdh.com
newsmax.com
newstamago.com
newstapa.org
newstarnet.com
newstatesman.com
newsweek.com
newtaiwan.com.tw
newtalk.tw
newthuhole.com
newyorker.com
newyorktimes.com
nexon.com
next11.co.jp
nextdigital.com.hk
nextmag.com.tw
nextmedia.com
nexton-net.jp
nexttv.com.tw
nf.id.au
nfjtyd.com
nflxext.com
nflximg.com
nflximg.net
nflxso.net
nflxvideo.net
nftstorage.link
nga.mil
ngensis.com
nhentai.net
nhi.gov.tw
nhk-ondemand.jp
nic.cz.cc
nic.gov
nicovideo.jp
nightswatch.top
nikke-en.com
nikke-jp.com
nikke-kr.com
nikke.hotcool.tw
nikkei.com
ninecommentaries.com
ninjacloak.com
ninjaproxy.ninja
nintendium.com
nirsoft.net
nitter.cc
nitter.net
niu.moe
niusnews.com
njactb.org
nko.navy.mil
nlfreevpn.com
nmsl.website
nnews.eu
no-ip.org
nobel.se
nobelprize.org
nodeseek.com
nokogiri.org
nokola.com
noodlevpn.com
norbulingka.org
nordstrom.com
nordstromimage.com
nordstromrack.com
nordvpn.com
nos.nl
notepad-plus-plus.org
nottinghampost.com
now.com
now.im
nownews.com
nowtorrents.com
noxinfluencer.com
npa.go.jp
npa.gov.tw
npm.gov.tw
npnt.me
npsboost.com
nradio.me
nrk.no
ns01.biz
ns01.info
ns01.us
ns02.biz
ns02.info
ns02.us
ns1.name
ns2.name
ns3.name
nsc.gov.tw
ntbk.gov.tw
ntbna.gov.tw
ntbt.gov.tw
ntd.tv
ntdtv.ca
ntdtv.co.kr
ntdtv.com
ntdtv.com.tw
ntdtv.cz
ntdtv.org
ntdtv.ru
ntdtvla.com
ntrfun.com
nubiles.net
nuexpo.com
nukistream.com
nurgo-software.com
nusatrip.com
nutaku.net
nutsvpn.work
nuuvem.com
nuvid.com
nvdst.com
nvquan.org
nvtongzhisheng.org
nwtca.org
nyaa.eu
nyaa.si
nybooks.com
nylon-angel.com
nylonstockingsonline.com
nypost.com
nyt.com
nytchina.com
nytcn.me
nytco.com
nyti.ms
nytimes.com
nytimes.map.fastly.net
nytimg.com
nytstyle.com
nzchinese.com
o3o.ca
oanda.com
oann.com
observechina.net
obutu.com
obyte.org
ocaspro.com
occupytiananmen.com
ocreampies.com
ocry.com
october-review.org
oculus.com
oculuscdn.com
odysee.com
oex.com
officeoftibet.com
ofile.org
ogaoga.org
ogate.org
ohmyrss.com
oikos.com.tw
oiktv.com
ok.ru
okayfreedom.com
okex.com
okk.tw
okpool.me
okx.com
old.honeynet.org
olehdtv.com
olelive.com
olevod.com
olumpo.com
olympicwatch.org
omct.org
omgili.com
omni7.jp
omnitalk.com
omnitalk.org
omny.fm
on.cc
on2.com
onapp.com
one.one.one.one
onedrive.com
onedrive.live.com
onedumb.com
onejav.com
onevps.com
onion.city
onion.ly
online.recoveryversion.org
onlinecha.com
onlineyoutube.com
onlygayvideo.com
onlytweets.com
onmoon.com
onmoon.net
onmypc.biz
onmypc.info
onmypc.net
onmypc.org
onthehunt.com
ontrac.com
oojj.de
open.com.hk
open.firstory.me
openai.com
opendemocracy.net
opendn.xyz
openid.net
openleaks.org
opensea.io
openstreetmap.org
opentech.fund
openvpn.net
openvpn.org
openwebster.com
openwrt.org.cn
opml.radiotime.com
opus-gaming.com
ordns.he.net
organcare.org.tw
organharvestinvestigation.net
organiccrap.com
orgasm.com
orgfree.com
oricon.co.jp
orient-doll.com
orientaldaily.com.my
orn.jp
osfoora.com
otcbtc.com
otto.de
otzo.com
ourdearamy.com
ourhobby.com
oursogo.com
oursteps.com.au
oursweb.net
ourtv.hk
overcast.fm
overdaily.org
overplay.net
oversea.istarshine.com
ovpn.com
ow.ly
owl.li
owltail.com
oxfordscholarship.com
oyax.com
oyghan.com
ozchinese.com
ozvoice.org
ozxw.com
ozyoyo.com
pachosting.com
pacificpoker.com
packetix.net
pacom.mil
pacopacomama.com
padmanet.com
page.link
page2rss.com
paimon.moe
palacemoon.com
paljorpublications.com
paltalk.com
panamapapers.sueddeutsche.de
pancakeswap.finance
pandafan.pub
pandapow.co
pandapow.net
pandavpn-jp.com
pandavpnpro.com
pandora.com
pandora.tv
panoramio.com
pao-pao.net
paper.li
paperb.us
paradisehill.cc
paradisepoker.com
parkansky.com
parler.com
parse.com
parsevideo.com
partycasino.com
partypoker.com
passion.com
passiontimes.hk
paste.ee
pastebin.com
pastie.org
patreon.com
patreonusercontent.com
pawoo.net
paxful.com
payments-jp.amazon.com
pbs.org
pbwiki.com
pbworks.com
pbxes.com
pbxes.org
pcanywhere.net
pcc.gov.tw
pcdvd.com.tw
pcgamestorrents.com
pchome.com.tw
pcij.org
pcstore.com.tw
pct.org.tw
pdetails.com
pdproxy.com
pds.nasa.gov
peace.ca
peacefire.org
peeasian.com
peing.net
pekingduck.org
pemulihan.or.id
pen.io
penchinese.com
pendrivelinux.com
penthouse.com
pentoy.hk
peoplebookcafe.com
peoplenews.tw
peopo.org
percy.in
perfect-privacy.com
perfectgirls.net
periscope.tv
perplexity.ai
persecutionblog.com
persiankitty.com
pewresearch.org
phapluan.org
phayul.com
philborges.com
phmsociety.org
phncdn.com
phosphation13.rssing.com
photodharma.net
photofocus.com
photonmedia.net
piaotia.com
picacomic.com
picacomiccn.com
picasaweb.com
picidae.net
picturedip.com
picturesocial.com
picuki.com
pigav.com
pimg.tw
pin-cong.com
pin6.com
pincong.rocks
ping.fm
pinimg.com
pinkrod.com
pinoy-n.com
pioneer-worker.forums-free.com
pipii.tv
piraattilahti.org
piring.com
pixeldrain.com
pixelqi.com
pixiv.net
pixiv.org
pixivsketch.net
pixnet.net
pk.com
pki.goog
pkqjiasu.com
pkuanvil.com
placemix.com
play-asia.com
playboy.com
playboyplus.com
player.fm
playno1.com
playpcesor.com
plexvpn.pro
plm.org.hk
plunder.com
plurk.com
plus.codes
plus28.com
plusbb.com
pmatehunter.com
pmates.com
po2b.com
pobieramy.top
podbean.com
podcast.co
podictionary.com
poe.com
points-media.com
pokerstars.com
pokerstars.net
politicalchina.org
politiscales.net
poloniex.com
polymarket.com
polymer-project.org
polymerhk.com
poolbinance.com
poolin.com
popai.pro
popo.tw
popvote.hk
popxi.click
popyard.com
popyard.org
porn.com
porn2.com
porn5.com
pornbase.org
pornerbros.com
pornhd.com
pornhost.com
pornhub.com
pornhubdeutsch.net
pornmate.com
pornoxo.com
pornrapidshare.com
pornsharing.com
pornsocket.com
pornstarbyface.com
pornstarclub.com
porntube.com
porntubenews.com
porntvblog.com
pornvisit.com
port25.biz
portablevpn.nl
poskotanews.com
post01.com
post76.com
post852.com
postadult.com
posts.careerengine.us
potato.im
potvpn.com
pourquoi.tw
power.com
powercx.com
powerphoto.org
ppy.sh
prayforchina.net
prcleader.org
premproxy.com
presentation.new
presentationzen.com
president.ir
presidentlee.tw
prestige-av.com
primevideo.com
prism-break.org
prisoneralert.com
pritunl.com
privacybox.de
privacyguides.org
private.com
privateinternetaccess.com
privatepaste.com
privatetunnel.com
privatevpn.com
privoxy.org
procopytips.com
project-syndicate.org
prosiben.de
proton.me
protonvpn.com
provideocoalition.com
provpnaccounts.com
proxfree.com
proxifier.com
proxomitron.info
proxpn.com
proxyanonimo.es
proxydns.com
proxynetwork.org.uk
proxyroad.com
proxytunnel.net
proxz.com
proyectoclubes.com
pscp.tv
pshvpn.com
psiphon.ca
psiphon3.com
psiphontoday.com
pstatic.net
pt.im
ptt.cc
pttgame.com
pttvan.org
ptwxz.com
pubu.com.tw
puffinbrowser.com
puffstore.com
pullfolio.com
punyu.com
pure18.com
pureconcepts.net
puredns.org
pureinsight.org
purepdf.com
purevpn.com
purplelotus.org
purpose.nike.com
pursuestar.com
pussyspace.com
putihome.org
putlocker.com
putty.org
pwned.com
pximg.net
python.com
python.com.tw
pythonhackers.com
pytorch.org
qbittorrent.org
qgirl.com.tw
qhigh.com
qi-gong.me
qianbai.tw
qiandao.today
qianglie.com
qiangwaikan.com
qiangyou.org
qianmo.tw
qidian.ca
qiwen.lu
qixianglu.cn
qkshare.com
qmp4.com
qoos.com
qpoe.com
qq.co.za
qstatus.com
qtrac.eu
questvisual.com
quitccp.org
quiz.directory
quora.com
quoracdn.net
quran.com
quranexplorer.com
qusi8.net
qxbbs.org
qz.com
r-pool.net
r0.ru
r10s.jp
r18.com
radicalparty.org
radiko.jp
radio-canada.ca
radio-en-ligne.fr
radio.garden
radioaustralia.net.au
radiohilight.net
radioline.co
radiovaticana.org
radiovncr.com
radmin-vpn.com
rael.org
raggedbanner.com
raidcall.com.tw
rainbowplan.org
raindrop.io
raizoji.or.jp
rakuten.co.jp
ramcity.com.au
rangzen.net
rangzen.org
rapbull.net
rapidmoviez.com
rapidvpn.com
rarbgprx.org
rateyourmusic.com
rationalwiki.org
raw.githack.com
rawgit.com
rawgithub.com
rcam.target.com
rcinet.ca
rconversation.blogs.com
rd.com
reabble.com
read01.com
read100.com
readingtimes.com.tw
readmoo.com
readydown.com
realcourage.org
realforum.zkiz.com
realitykings.com
realraptalk.com
realsexpass.com
reason.com
rebatesrule.net
recordhistory.org
recovery.org.tw
recoveryversion.com.tw
red-lang.org
redbubble.com
redchinacn.net
redchinacn.org
redd.it
reddit.com
reddithelp.com
redditlist.com
redditmedia.com
redditspace.com
redditstatic.com
redhotlabs.com
redtube.com
referer.us
reflectivecode.com
relaxbbs.com
relay.com.tw
relay.firefox.com
releaseinternational.org
religionnews.com
renminbao.com
renyurenquan.org
resilio.com
resistchina.org
retweetist.com
retweetrank.com
reuters.com
reutersmedia.net
revleft.com
revver.com
rfa.org
rfachina.com
rfalive1.akacast.akamaistream.net
rfamobile.org
rfaweb.org
rferl.org
rfi.fr
rfi.my
rigpa.org
riku.me
rileyguide.com
riseup.net
ritouki.jp
ritter.vg
rixcloud.com
rixcloud.us
rlwlw.com
rmbl.ws
rmjdw.com
roadshow.hk
roboforex.com
robustnessiskey.com
rocket-inc.net
rocket.chat
rocksdb.org
rojo.com
rolfoundation.org
rolia.net
rolsociety.org
ronjoneswriter.com
roodo.com
rosechina.net
rou.video
rsdlmonitor.com
rsf-chinese.org
rsf.org
rsshub.app
rssmeme.com
rtalabel.org
rthk.hk
rthk.org.hk
rthklive2-lh.akamaihd.net
rti.org.tw
rti.tw
rtm.tnt-ea.com
ruanyifeng.com
rukor.org
rule34.xxx
rule34video.com
rumble.com
runbtx.com
rushbee.com
rusvpn.com
ruten.com.tw
rutracker.net
rutracker.org
rutube.ru
rxhj.net
s-cute.com
s-dragon.org
s.yimg.com
s1s1s1.com
s3-ap-northeast-1.amazonaws.com
s3-ap-northeast-2.amazonaws.com
s3-ap-southeast-1.amazonaws.com
s3-ap-southeast-2.amazonaws.com
s3-eu-central-1.amazonaws.com
s3.amazonaws.com
s3.ap-northeast-2.amazonaws.com
s3.eu-central-1.amazonaws.com
s3.us-east-1.amazonaws.com
sa.hao123.com
sacks.com
sacom.hk
sadistic-v.com
sadpanda.us
safechat.com
safeguarddefenders.com
safervpn.com
sagernet.org
saintyculture.com
sakuralive.com
sakya.org
salvation.org.hk
samair.ru
sambhota.org
sankakucomplex.com
sankei.com
sanmin.com.tw
sapikachu.net
savemedia.com
savethedate.foo
savethesounds.info
savetibet.de
savetibet.fr
savetibet.nl
savetibet.org
savetibet.ru
savetibetstore.org
saveuighur.org
savevid.com
sbme.me
sbs.com.au
scache.vzw.com
scache1.vzw.com
scache2.vzw.com
scasino.com
schema.org
sciencenets.com
scieron.com
scmp.com
scmpchinese.com
scramble.io
scratch.mit.edu
scribd.com
scriptspot.com
search.aol.com
search.com
search.xxx
search.yahoo.co.jp
searchtruth.com
searx.me
seattlefdc.com
secretchina.com
secretgarden.no
secretsline.biz
secure.hustler.com
secure.logmein.com
secure.shadowsocks.nu
secureservercdn.net
securetunnel.com
securityinabox.org
securitykiss.com
seed4.me
seesmic.com
seevpn.com
seezone.net
sehuatang.net
sehuatang.org
sejie.com
sellclassics.com
sendsmtp.com
sendspace.com
sensortower.com
servehttp.com
serveuser.com
serveusers.com
sesawe.net
sethwklein.net
setn.com
settv.com.tw
sevenload.com
sex.com
sex3.com
sex8.cc
sexandsubmission.com
sexbot.com
sexhu.com
sexidude.com
sexinsex.net
sextvx.com
sf.net
sfileydy.com
sfshibao.com
sftindia.org
sftuk.org
shadeyouvpn.com
shadow.ma
shadowsky.xyz
shadowsocks.asia
shadowsocks.be
shadowsocks.com
shadowsocks.com.hk
shadowsocks.org
shahit.biz
shambalapost.com
shapeservices.com
share-videos.se
share.america.gov
share.ovi.com
share.youthwant.com.tw
sharebee.com
sharecool.org
sharpdaily.hk
sharpdaily.tw
shat-tibet.com
shattered.io
sheet.new
sheets.new
sheikyermami.com
shellfire.de
shenyun.com
shenyunperformingarts.org
shenyunshop.com
shenzhoufilm.com
shenzhouzhengdao.org
shiatv.net
shicheng.org
shiksha.com
shipcamouflage.com
shireyishunjian.com
shitaotv.org
shixiao.org
shizhao.org
shkspr.mobi
shodanhq.com
shooshtime.com
shop2000.com.tw
shopee.tw
shopping.com
shopping.yahoo.co.jp
showhaotu.com
showtime.jp
showwe.tw
shutterstock.com
shwchurch.org
shwchurch3.com
siddharthasintent.org
sidelinesnews.com
sidelinessportseatery.com
sierrafriendsoftibet.org
signal.org
sijihuisuo.club
sijihuisuo.com
silkbook.com
silvergatebank.com
simbolostwitter.com
simplecd.me
simplecd.org
simpleproductivityblog.com
simpleswap.io
simplex.chat
sina.com.hk
sinchew.com.my
singaporepools.com.sg
singfortibet.com
singlelogin.se
singpao.com.hk
singtao.com
singtaousa.com
sino-monthly.com
sinoants.com
sinoca.com
sinocast.com
sinocism.com
sinoinsider.com
sinomontreal.ca
sinoquebec.com
sipml5.org
sis.xxx
sis001.com
sis001.us
site.new
site2unblock.com
sitebro.tw
sitekreator.com
sitemaps.org
sites.new
six-degrees.io
sketchappsources.com
skimtube.com
skybet.com
skyking.com.tw
skyvegas.com
skyxvpn.com
slacker.com
slashine.onl
slaytizle.com
sleazydream.com
sleazyfork.org
slheng.com
slickvpn.com
slides.com
slides.new
slideshare.net
slinkset.com
slutload.com
slutmoonbeam.com
slyip.com
slyip.net
sm-miracle.com
smartdnsproxy.com
smarthide.com
smchbooks.com
smh.com.au
smhric.org
smith.edu
smn.news
smyxy.org
snapseed.com
sndcdn.com
sneakme.net
snowlionpub.com
sobees.com
soc.mil
social.datalabour.com
social.edu.ci
socialblade.com
socks-proxy.net
sockscap64.com
sockslist.net
socrec.org
sod.co.jp
softether-download.com
softether.co.jp
softether.org
softfamous.com
softwarebychuck.com
softwaredownload.gitbooks.io
sogclub.com
sogrady.me
soh.tw
sohcradio.com
sohfrance.org
sokamonline.com
sokmil.com
solana.com
solarsystem.nasa.gov
solidaritetibet.org
solidfiles.com
solv.finance
somee.com
songjianjun.com
sonidodelaesperanza.org
sopcast.com
sopcast.org
sorting-algorithms.com
sos.org
sosad.fun
sosreader.com
soubory.com
soul-plus.net
soulcaliburhentai.net
soundcloud.com
soundofhope.kr
soundofhope.org
soundon.fm
soup.io
sourceforge.net
sourcewadio.com
south-plus.net
south-plus.org
southmongolia.org
southnews.com.tw
southpark.cc.com
sowers.org.hk
soylentnews.org
spaces.hightail.com
spankbang.com
spankingtube.com
spankwire.com
sparkpool.com
spatial.io
spb.com
speakerdeck.com
specxinzl.jigsy.com
speedcat.me
speedify.com
spencertipping.com
spendee.com
spicevpn.com
spideroak.com
spiderpool.com
spike.com
sports.williamhill.com
spotflux.com
spotify.com
spreadsheet.new
spreadshirt.es
spreaker.com
spring4u.info
springboardplatform.com
springwood.me
sprite.org
sproutcore.com
squirly.info
squirrelvpn.com
srcf.ucam.org
ss-link.com
ss.pythonic.life
ss7.vzw.com
ssglobal.co
ssglobal.me
ssl.webpack.de
ssl443.org
sspanel.net
ssr.tools
ssrshare.com
ssrshare.us
ssrtool.com
sstm.moe
sstmlt.moe
sstmlt.net
stackoverflow.com
standard.co.uk
standupfortibet.org
standwithhk.org
stanford.edu
starfishfx.com
starp2p.com
startpage.com
startuplivingchina.com
stat.gov.tw
static-economist.com
static.shemalez.com
static01.nyt.com
staticflickr.com
stboy.net
stc.com.sa
steamcommunity.com
steamstatic.com
steel-storm.com
steemit.com
steganos.com
steganos.net
stepchina.com
stephaniered.com
sthoo.com
stickam.com
stickeraction.com
stileproject.com
stitcher.com
sto.cc
stoporganharvesting.org
stoptibetcrisis.net
storage.yandex.net
storagenewsletter.com
store.steampowered.com
storj.io
storm.mg
stormmediagroup.com
storry.tv
stoweboyd.com
straitstimes.com
stranabg.com
straplessdildo.com
streamable.com
streamate.com
streamingthe.net
streema.com
strikingly.com
strongvpn.com
strongwindpress.com
studentsforafreetibet.org
stumbleupon.com
stupidvideos.com
subhd.tv
substack.com
successfn.com
suche.gmx.net
sugarsync.com
sugobbs.com
sugumiru18.com
suissl.com
sujiatun.wordpress.com
summify.com
sumrando.com
sun1911.com
sundayguardianlive.com
sunmedia.ca
suno.ai
suno.com
sunporno.com
sunskyforum.com
sunta.com.tw
sunvpn.net
sunwinism.joinbbs.net
supchina.com
superfreevpn.com
superpages.com
supervpn.net
superzooi.com
suppig.net
suprememastertv.com
surfeasy.com
surfeasy.com.au
surfshark.com
suroot.com
surrenderat20.net
svsfx.com
swagbucks.com
swapspace.co
swissinfo.ch
swissvpn.net
switch1.jp
switchvpn.net
sydneytoday.com
sylfoundation.org
synapse.org
syncback.com
synergyse.com
syosetu.com
sysresccd.org
sytes.net
szbbs.net
szetowah.org.hk
t-g.com
t.co
t.me
t35.com
t66y.com
taa-usa.org
taaze.tw
tabtter.jp
tacc.cwb.gov.tw
taconet.com.tw
taedp.org.tw
tafm.org
tagwalk.com
tahr.org.tw
taipei.gov.tw
taipeisociety.org
taipeitimes.com
taisounds.com
taiwanbible.com
taiwandaily.net
taiwandc.org
taiwanhot.net
taiwanjobs.gov.tw
taiwanjustice.com
taiwanjustice.net
taiwankiss.com
taiwannation.50webs.com
taiwannation.com
taiwannation.com.tw
taiwanncf.org.tw
taiwannews.com.tw
taiwantp.net
taiwantt.org.tw
taiwanus.net
taiwanyes.ning.com
talk853.com
talkatone.com
talkboxapp.com
talkcc.com
talkonly.net
tanc.org
tangren.us
tanks.gg
taoism.net
tapanwap.com
tapatalk.com
tardigrade.io
tarr.uspto.gov
taup.net
taweet.com
tbcollege.org
tbi.org.hk
tbjyt.org
tbrc.org
tbs-rainbow.org
tbsec.org
tbskkinabalu.page.tl
tbsn.org
tbsseattle.org
tbssqh.org
tbswd.org
tbtemple.org.uk
tbthouston.org
tccwonline.org
tcewf.org
tchrd.org
tcnynj.org
tcpspeed.co
tcsofbc.org
tdm.com.mo
teachparentstech.org
teamamericany.com
technews.tw
techspot.com
techviz.net
teck.in
teco-hk.org
teco-mo.org
teddysun.com
teeniefuck.net
teensinasia.com
tehrantimes.com
telecomspace.com
telega.one
telegra.ph
telegram.dog
telegram.me
telegram.org
telegram.space
telegramdownload.com
telegraph.co.uk
telesco.pe
tellapart.com
tellme.pw
tenacy.com
tenor.com
tensorflow.org
tenzinpalmo.com
terabox.com
tew.org
textnow.com
textnow.me
tfc-taiwan.org.tw
tfhub.dev
tfiflve.com
tg-me.com
tg.dev
th.hao123.com
thaicn.com
thb.gov.tw
theatlantic.com
theatrum-belli.com
thebcomplex.com
theblaze.com
theblemish.com
thebobs.com
thebodyshop-usa.com
thecenter.mit.edu
thechasernews.co.uk
thechinabeat.org
thechinacollection.org
theconversation.com
thedalailamamovie.com
thediplomat.com
thedw.us
theepochtimes.com
thefacebook.com
thegay.com
thegioitinhoc.vn
thegly.com
theguardian.com
thehansindia.com
thehindu.com
thehun.net
theinitium.com
thenewslens.com
thepiratebay.org
theporndude.com
theportalwiki.com
theprint.in
therock.net.nz
thesaturdaypaper.com.au
thestandnews.com
thetatoken.org
thetibetcenter.org
thetibetconnection.org
thetibetmuseum.org
thetibetpost.com
thetrotskymovie.com
thetvdb.com
thewgo.org
thewirechina.com
theync.com
thinkgeek.com
thinkingtaiwan.com
thinkwithgoogle.com
thirdmill.org
thisav.com
thlib.org
thomasbernhard.org
thongdreams.com
threadreaderapp.com
threads.com
threads.net
throughnightsfire.com
thuhole.com
thumbzilla.com
thywords.com
tiananmenduizhi.com
tiananmenmother.org
tiananmenuniv.com
tiananmenuniv.net
tiandixing.org
tianhuayuan.com
tianlawoffice.com
tianti.io
tiantibooks.org
tianyantong.org.cn
tianzhu.org
tibet-envoy.eu
tibet-foundation.org
tibet-house-trust.co.uk
tibet-initiative.de
tibet-munich.de
tibet.at
tibet.ca
tibet.com
tibet.fr
tibet.net
tibet.nu
tibet.org
tibet.org.tw
tibet.to
tibet3rdpole.org
tibetaction.net
tibetaid.org
tibetalk.com
tibetan-alliance.org
tibetan.fr
tibetanaidproject.org
tibetanarts.org
tibetanbuddhistinstitute.org
tibetancommunity.org
tibetancommunityuk.net
tibetanculture.org
tibetanentrepreneurs.org
tibetanfeministcollective.org
tibetanhealth.org
tibetanjournal.com
tibetanlanguage.org
tibetanliberation.org
tibetanpaintings.com
tibetanphotoproject.com
tibetanpoliticalreview.org
tibetanreview.net
tibetansports.org
tibetanwomen.org
tibetanyouth.org
tibetanyouthcongress.org
tibetcharity.dk
tibetcharity.in
tibetchild.org
tibetcity.com
tibetcollection.com
tibetcorps.org
tibetexpress.net
tibetfocus.com
tibetfund.org
tibetgermany.com
tibetgermany.de
tibethaus.com
tibetheritagefund.org
tibethouse.jp
tibethouse.org
tibethouse.us
tibetinfonet.net
tibetjustice.org
tibetkomite.dk
tibetlibre.free.fr
tibetmuseum.org
tibetnetwork.org
tibetoffice.ch
tibetoffice.com.au
tibetoffice.eu
tibetoffice.org
tibetonline.com
tibetonline.tv
tibetoralhistory.org
tibetpolicy.eu
tibetrelieffund.co.uk
tibetsociety.com
tibetsun.com
tibetsupportgroup.org
tibetswiss.ch
tibettelegraph.com
tibettimes.net
tibettruth.com
tibetwrites.org
ticket.com.tw
tigervpn.com
tiktok.com
tiktokcdn-eu.com
tiktokcdn-us.com
tiktokcdn.com
tiktokv.com
tiktokv.us
tiltbrush.com
timdir.com
time.com
timesnownews.com
timesofindia.indiatimes.com
timsah.com
timtales.com
tinc-vpn.org
tineye.com
tingtalk.me
tiny.cc
tinychat.com
tinypaste.com
tinyurl.com
tipas.net
tipo.gov.tw
tistory.com
tkcs-collins.com
tl.gd
tma.co.jp
tmagazine.com
tmi.me
tmpp.org
tn1.shemalez.com
tn2.shemalez.com
tn3.shemalez.com
tnaflix.com
tnp.org
to-porno.com
togetter.com
toh.info
token.im
tokenlon.im
tokyo-247.com
tokyo-hot.com
tokyo-porn-tube.com
tokyocn.com
tomp3.cc
tongil.or.kr
tonyyan.net
toonel.net
top.tv
top10vpn.com
top81.ws
topbtc.com
topic.youthwant.com.tw
topnews.in
toppornsites.com
topshareware.com
topsy.com
toptip.ca
toptoon.net
tor.updatestar.com
tora.to
torcn.com
torguard.net
torlock.com
torproject.org
torrentgalaxy.to
torrentkitty.tv
torrentprivacy.com
torrentproject.se
torrenty.org
tortoisesvn.net
torvpn.com
tosh.comedycentral.com
totalvpn.com
tou.tv
toutiaoabc.com
towngain.com
toypark.in
toythieves.com
toytractorshow.com
tparents.org
tpi.org.tw
tracfone.com
tradingview.com
translate.goog
transparency.org
treemall.com.tw
trendsmap.com
trickip.net
trimondi.de
tronscan.org
trouw.nl
trt.net.tr
trtc.com.tw
truebuddha-md.org
trulyergonomic.com
truthsocial.com
truveo.com
tryheart.jp
tsctv.net
tsdr.uspto.gov
tsemtulku.com
tsquare.tv
tsu.org.tw
tsunagarumon.com
tt1069.com
tttan.com
ttv.com.tw
ttvnw.net
tu8964.com
tubaholic.com
tube.com
tube8.com
tube911.com
tubecup.com
tubegals.com
tubeislam.com
tubepornclassic.com
tubestack.com
tubewolf.com
tuibeitu.net
tuidang.org
tuidang.se
tuitwit.com
tukaani.org
tumblr.com
tumutanzi.com
tumview.com
tunein.com
tunein.streamguys1.com
tunnelbear.com
tunnelblick.net
tunnelr.com
tunsafe.com
turansam.org
turbobit.net
turbohide.com
turkistantimes.com
turntable.fm
tushycash.com
tuvpn.com
tuzaijidi.com
tv.com
tv.jtbc.joins.com
tvants.com
tvboxnow.com
tvider.com
tvmost.com.hk
tvplayvideos.com
tvunetworks.com
tw-blog.com
tw-npo.org
tw.gigacircle.com
tw.hao123.com
tw.jiepang.com
tw.streetvoice.com
tw.tomonews.net
tw01.org
twaitter.com
twapperkeeper.com
twaud.io
twavi.com
twbbs.org
twblogger.com
tweepguide.com
tweepmag.com
tweepml.org
tweetbackup.com
tweetboard.com
tweetcs.com
tweetdeck.com
tweetedtimes.com
tweetphoto.com
tweetree.com
tweettunnel.com
tweetwally.com
tweetymail.com
tweez.net
twelve.today
twerkingbutt.com
twftp.org
twgreatdaily.com
twibase.com
twibble.de
twibbon.com
twibs.com
twicountry.org
twicsy.com
twiends.com
twifan.com
twiffo.com
twiggit.org
twilightsex.com
twilog.org
twimbow.com
twimg.com
twimg.edgesuite.net
twip.me
twipple.jp
twishort.com
twister.net.co
twisternow.com
twistory.net
twitch.tv
twitchcdn.net
twitgoo.com
twitiq.com
twitlonger.com
twitmania.com
twitoaster.com
twitonmsn.com
twitpic.com
twitstat.com
twittbot.net
twitter.com
twitter.jp
twitter4j.org
twittercounter.com
twitterfeed.com
twittergadget.com
twitterkr.com
twittermail.com
twitterrific.com
twittertim.es
twitthat.com
twitturk.com
twitturly.com
twitzap.com
twiyia.com
twkan.com
twnorth.org.tw
twreporter.org
twskype.com
twt.tl
twtkr.com
twtr2src.ogaoga.org
twtrland.com
twttr.com
twurl.nl
tx.me
txxx.com
tycool.com
typepad.com
typeset.io
typora.io
u15.info
u9un.com
ua5v.com
ub0.cc
ubddns.org
uberproxy.net
uc-japan.org
uchicago.edu
udn.com
udn.com.tw
udnbkk.com
udomain.hk
uforadio.com.tw
ufreevpn.com
ugo.com
uhdwallpapers.org
uhrp.org
uighur.narod.ru
uighur.nl
uighurbiz.net
ukcdp.co.uk
uku.im
ulike.net
ulop.net
ultrasurf.us
ultravpn.com
ultravpn.fr
ultraxs.com
umich.edu
unblock-us.com
unblock.cn.com
unblockdmm.com
unblocker.yt
unblocksit.es
uncyclomedia.org
uncyclopedia.hk
uncyclopedia.tw
underwoodammo.com
unholyknight.com
uni.cc
unification.net
unification.org.tw
unirule.cloud
unix100.com
unknownspace.org
unmineable.com
unodedos.com
unpo.org
unseen.is
unstable.icu
unwire.hk
uocn.org
upbit.com
updates.tdesktop.com
upghsbc.com
upholdjustice.org
uploaded.net
uploaded.to
uploadstation.com
upmedia.mg
upornia.com
uproxy.org
uptodown.com
upwill.org
ur7s.com
uraban.me
urbandictionary.com
urbansurvival.com
urchin.com
urlborg.com
urlparser.com
us.to
usacn.com
usaip.eu
uscardforum.com
uscg.mil
uscnpm.org
use.typekit.net
usercontent.goog
users.skynet.be
usfk.mil
usma.edu
usmgtcg.ning.com
usno.navy.mil
usocctn.com
ustibetcommittee.org
ustream.tv
usus.cc
utopianpal.com
uujiasu.com
uukanshu.com
uupool.cn
uvwxyz.xyz
uwants.com
uwants.net
uyghur-j.org
uyghur.co.uk
uyghuraa.org
uyghuramerican.org
uyghurbiz.org
uyghurcongress.org
uyghurpen.org
uyghurstudies.org
uyghurtribunal.com
uygur.fc2web.com
uygur.org
uymaarip.com
v2.help
v2ex.com
v2fly.org
v2ray.com
v2raycn.com
valeursactuelles.com
van001.com
van698.com
vanemu.cn
vanilla-jp.com
vanpeople.com
vansky.com
vaticannews.va
vatn.org
vcf-online.org
vcfbuilder.org
vegas.williamhill.com
vegasred.com
velkaepocha.sk
venbbs.com
venchina.com
venetianmacao.com
ventureswell.com
veoh.com
vercel.app
vermonttibet.org
vern.cc
verybs.com
vevo.com
vewas.net
vft.com.tw
viber.com
vica.info
vid.me
vidble.com
video.aol.ca
video.aol.co.uk
video.aol.com
video.foxbusiness.com
videobam.com
videodetective.com
videomega.tv
videomo.com
videopediaworld.com
videopress.com
vidinfo.org
vietdaikynguyen.com
vijayatemple.org
vilanet.me
vilavpn.com
vimeo.com
vimperator.org
vincnd.com
vine.co
vinniev.com
vip-enterprise.com
virtualrealporn.com
visibletweets.com
viu.com
viu.tv
vivahentai4u.net
vivaldi.com
vivatube.com
vivthomas.com
vizvaz.com
vjav.com
vjmedia.com.hk
vllcs.org
vmixcore.com
vmpsoft.com
vn.hao123.com
vnet.link
voa-11.akacast.akamaistream.net
voacambodia.com
voacantonese.com
voachinese.com
voachineseblog.com
voagd.com
voaindonesia.com
voanews.com
voatibetan.com
voatibetanenglish.com
vocaroo.com
vocativ.com
vocn.tv
vocus.cc
vod-abematv.akamaized.net
vod.wwe.com
voicettank.org
vot.org
vovo2000.com
voxer.com
voy.com
vpl.bibliocommons.com
vpn.ac
vpn.cmu.edu
vpn.net
vpn.sv.cmu.edu
vpn4all.com
vpnaccount.org
vpnaccounts.com
vpnbook.com
vpncomparison.org
vpncoupons.com
vpncup.com
vpndada.com
vpnfan.com
vpnfire.com
vpnforgame.net
vpngate.jp
vpngate.net
vpngratis.net
vpnhq.com
vpnhub.com
vpninja.net
vpnintouch.com
vpnjack.com
vpnmaster.com
vpnmentor.com
vpnpick.com
vpnpop.com
vpnpronet.com
vpnproxymaster.com
vpnreactor.com
vpnreviewz.com
vpnsecure.me
vpnshazam.com
vpnshieldapp.com
vpnsp.com
vpntraffic.com
vpntunnel.com
vpnuk.info
vpnunlimitedapp.com
vpnvip.com
vpnworldwide.com
vporn.com
vpser.net
vpsxb.net
vraiesagesse.net
vrchat.com
vrmtr.com
vrporn.com
vrsmash.com
vtunnel.com
vuku.cc
vultryhw.com
w-pool.com
w.idaiwan.com
w3s.link
waffle1999.com
wahas.com
waikeung.org
wainao.me
walletconnect.com
wallmama.com
wallpapercasa.com
wallproxy.com
wallsttv.com
waltermartin.com
waltermartin.org
wanderinghorse.net
wangafu.net
wangjinbo.org
wanglixiong.com
wango.org
wangruoshui.net
want-daily.com
wanz-factory.com
wapedia.mobi
warroom.org
waselpro.com
washingtonpost.com
watch8x.com
watchinese.com
watchmygf.net
watchout.tw
wattpad.com
wav.tv
waveprotocol.org
waybig.com
waymo.com
wd.bible
wda.gov.tw
wdf5.com
wealth.com.tw
wearehairy.com
wearn.com
web.dev
web2project.net
webbang.net
webevader.org
webfreer.com
webjb.org
weblagu.com
webmproject.org
webpkgcache.com
webrtc.org
webrush.net
webs-tv.net
website.informer.com
website.new
websitepulse.com
webwarper.net
webworkerdaily.com
wechatlawsuit.com
weebly.com
wefightcensorship.org
wefong.com
wego.here.com
weiboleak.com
weihuo.org
weijingsheng.org
weiming.info
weiquanwang.org
weisuo.ws
welovecock.com
welt.de
wemigrate.org
wengewang.com
wengewang.org
wenxuecity.com
wenyunchao.com
wenzhao.ca
westca.com
westernshugdensociety.org
westernwolves.com
westkit.net
westpoint.edu
wetplace.com
wetpussygames.com
wezone.net
wforum.com
whatblocked.com
whatbrowser.org
whats.new
whatsapp.com
whatsapp.net
wheelockslatin.com
whereiswerner.com
wheretowatch.com
whippedass.com
whispersystems.org
whoer.net
whotalking.com
whylover.com
whyx.org
widevine.com
wikaba.com
wiki.gamerp.jp
wiki.jqueryui.com
wiki.keso.cn
wiki.metacubex.one
wiki.oauth.net
wiki.phonegap.com
wikibooks.org
wikidata.org
wikileaks-forum.com
wikileaks.ch
wikileaks.com
wikileaks.de
wikileaks.eu
wikileaks.lu
wikileaks.org
wikileaks.pl
wikiless.funami.tech
wikilivres.info
wikimapia.org
wikimedia.org
wikinews.org
wikipedia.org
wikisource.org
wikiversity.org
wikivoyage.org
wikiwand.com
wiktionary.org
williamhill.com
willw.net
wilsoncenter.org
windscribe.com
wingamestore.com
wingy.site
winning11.com
wionews.com
wire.com
wiredbytes.com
wiredpen.com
wireguard.com
wisdompubs.org
wisevid.com
withgoogle.com
withyoutube.com
witnessleeteaching.com
witopia.net
wizcrafts.net
wjbk.org
wlcnew.jigsy.com
wmflabs.org
wmfusercontent.org
wn.com
wnacg.com
wnacg.org
wo.tc
wo3ttt.wordpress.com
woeser.com
wokar.org
wolfax.com
wombo.ai
woolyss.com
woopie.jp
woopie.tv
wordpress.com
work2icu.org
workatruna.com
workerempowerment.org
workers.dev
worldcat.org
worldjournal.com
worldvpn.net
wow.com
wowgirls.com
wowhead.com
wowporn.com
wowrk.com
woyaolian.org
wozy.in
wp.com
wpoforum.com
wrchina.org
wretch.cc
writer.zoho.com
writesonic.com
wsj.com
wsj.net
wtbn.org
wtfpeople.com
wuerkaixi.com
wufafangwen.com
wufi.org.tw
wujie.net
wujieliulan.com
wunderground.com
wuw.red
wwitv.com
www.ajsands.com
www.antd.org
www.aolnews.com
www.businessinsider.com.au
www.cmoinc.org
www.dmm.com
www.dwheeler.com
www.eastturkistan.net
www.gmiddle.com
www.gmiddle.net
www.hustlercash.com
www.idlcoyote.com
www.imdb.com
www.kindleren.com
www.klip.me
www.lib.virginia.edu
www.lorenzetti.com.br
www.m-sport.co.uk
www.monlamit.org
www.moztw.org
www.msn.com
www.nbc.com
www.owind.com
www.oxid.it
www.powerpointninja.com
www.s4miniarchive.com
www.sciencemag.org
www.shadowsocks.com
www.skype.com
www.tablesgenerator.com
www.taiwanonline.cc
www.thechinastory.org
www.wan-press.org
www.websnapr.com
www.xicons.org
www.zensur.freerk.com
www1.american.edu
www1.biz
www2.ohchr.org
www2.rocketbbs.com
wwwhost.biz
wxw.cat
wxw.moe
wzyboy.im
x-art.com
x-berry.com
x-wall.org
x.ai
x.co
x.com
x.company
x24hr.com
x3guide.com
xanga.com
xbabe.com
xbookcn.com
xbtce.com
xcafe.in
xcity.jp
xcritic.com
xerotica.com
xfinity.com
xfxssr.me
xgmyd.com
xhamster.com
xianba.net
xianjian.tw
xiaobaiwu.com
xiaochuncnjp.com
xiaohexie.com
xiaolan.me
xiaoma.org
xiaomi.eu
xiaxiaoqiang.net
xiezhua.com
xihua.es
xijie.wordpress.com
xing.com
xinjiangpolicefiles.org
xinmiao.com.hk
xinqimeng.over-blog.com
xinsheng.net
xinshijue.com
xiongpian.com
xiuren.org
xizang-zhiye.org
xjp.cc
xjtravelguide.com
xm.com
xml-training-guide.com
xmovies.com
xn--11xs86f.icu
xn--4gq171p.com
xn--9pr62r24a.com
xn--czq75pvv1aj5c.org
xn--i2ru8q2qg.com
xn--ngstr-lra8j.com
xn--noss43i.com
xn--oiq.cc
xn--p8j9a0d9c9a.xn--q9jyb4c
xnpool.com
xnxx.com
xpdo.net
xpud.org
xrentdvd.com
xsden.info
xskywalker.com
xt.com
xt.pub
xtube.com
xuchao.net
xuchao.org
xuehua.us
xvbelink.com
xvideo.cc
xvideos-cdn.com
xvideos.com
xvideos.es
xvinlink.com
xxbbx.com
xxlmovies.com
xxuz.com
xxx.com
xxx.xxx
xxxfuckmom.com
xxxx.com.au
xxxy.info
xxxymovies.com
xys.dxiong.com
xys.org
xysblogs.org
y2mate.com
yadi.sk
yahoo.com
yahoo.com.hk
yahoo.com.tw
yakbutterblues.com
yam.com
yam.org.tw
yande.re
yanghengjun.com
yangzhi.org
yasni.co.uk
yasukuni.or.jp
yayabay.com
ydy.com
yeahteentube.com
yecl.net
yeelou.com
yeeyi.com
yegle.net
yes-news.com
yes.xxx
yes123.com.tw
yesasia.com
yesasia.com.hk
yespornplease.com
yeyeclub.com
ygto.com
yhcw.net
yibada.com
yibaochina.com
yidio.com
yigeni.com
yilubbs.com
yinlei.org
yipub.com
yizhihongxing.com
yobit.net
yobt.com
yobt.tv
yogichen.org
yolasite.com
yomiuri.co.jp
yong.hu
yorkbbs.ca
you-get.org
you.com
youdontcare.com
youjizz.com
youmaker.com
youngpornvideos.com
youngspiration.hk
youpai.org
youporn.com
youporngay.com
your-freedom.net
yourepeat.com
yourlisten.com
yourlust.com
yourtrap.com
yousendit.com
youthnetradio.org
youtu.be
youtube-nocookie.com
youtube.com
youtubeeducation.com
youtubegaming.com
youtubekids.com
youversion.com
youwin.com
youxu.info
yt.be
ytht.net
ytimg.com
ytn.co.kr
yuanming.net
yuanzhengtang.org
yulghun.com
yunchao.net
yunomi.tokyo
yuvutu.com
yvesgeleyn.com
ywpw.com
yx51.net
yyii.org
yyjlymb.xyz
yysub.net
yzzk.com
z-lib.fm
z-lib.fo
z-lib.gd
z-lib.gl
z-lib.io
z-lib.org
z-library.sk
zacebook.com
zalmos.com
zamimg.com
zaobao.com.sg
zapto.org
zattoo.com
zb.com
zdnet.com.tw
zello.com
zengjinyan.org
zenmate.com
zenmate.com.ru
zerohedge.com
zeronet.io
zfreet.com
zh-hans.cfsh99.com
zh.ecdm.wikia.com
zh.pokerstrategy.com
zh.pttpedia.wikia.com
zh.uncyclopedia.wikia.com
zh.wikiquote.org
zhangboli.net
zhangtianliang.com
zhanlve.org
zhao.1984.city
zhao.jinhai.de
zhenghui.org
zhengjian.org
zhengwunet.org
zhenxiang.biz
zhizhu.top
zhongguo.ca
zhongguorenquan.org
zhongguotese.net
zhongzidi.com
zhoushuguang.com
zhuanxing.cn
zhuatieba.com
zhuichaguoji.org
zi.media
ziddu.com
zillionk.com
zim.vn
zinio.com
ziporn.com
zippyshare.com
zmedia.com.tw
zmw.cn
zodgame.us
zodgame.xyz
zomobo.net
zonaeuropa.com
zonghexinwen.com
zoogvpn.com
zoominfo.com
zooqle.com
zootool.com
zoozle.net
zophar.net
zorrovpn.com
zozotown.com
zpn.im
zspeeder.me
zsrhao.com
zuo.la
zuobiao.me
zuola.com
zvereff.com
zynamics.com
zyns.com
zyxel.com
zzcartoon.com
zzcloud.me
zzux.com
|
297854895/vue-tsx-admin | 3,284 | README.md | 在线示例
----
[在线预览](http://115.126.79.121:9999/)
页面示例完成情况
----
- [x] 登录页
- [x] 首页
- [ ] 表单页
- [ ] 基础表单
- [ ] 分步表单
- [ ] 高级表单
- [ ] 列表页
- [x] 查询表格
- [ ] 标准列表
- [ ] 卡片列表
- [ ] 搜索列表
- [ ] 详情页
- [ ] 标准详情
- [ ] 高级详情
- [ ] 个人页
- [ ] 个人中心
- [ ] 个人设置
- [ ] 开发引导示例
- [ ] 主题色
- [ ] Tab组件
- [ ] 整体风格
- [ ] 其他设置
- [ ] 路由动画
- [x] 数据持久化
- [x] 系统数据持久化
- [x] 路由数据持久化
- [ ] 权限管理
- [ ] 登录验证
- [ ] 创建路由
- [ ] 创建菜单
- [x] 404页
主体完成情况
----
- [x] 登录
- [ ] 注册
- [x] 退出登录
- [x] 登录验证
- [x] 权限创建菜单
- [x] 权限验证路由合法性
- [x] 系统样式自定义
- [x] 响应式布局
- [x] 数据持久化
- [x] 路由按需引入
- [x] 支撑浏览器前进后退按钮的Tab管理组件
- [ ] 同路由多开Tab
- [x] 国际化配置
- [ ] 完善说明文档
概述
----
基于`typescript`、`jsx`、`vue`、`ant-design-vue`的中后台模板


下载与运行
----
- 拉取项目
```bash
git clone https://github.com/297854895/vue-tsx-admin.git
cd vue-tsx-admin
```
- 安装依赖
```
yarn install
```
- 开发模式运行
```
yarn run serve
```
- 编译项目
```
yarn run build
```
- Lints and fixes files
```
yarn run lint
```
环境和依赖
----
- node
- webpack
- eslint
- @vue/cli ~3
- [ant-design-vue](https://github.com/vueComponent/ant-design-vue) - Ant Design Of Vue 实现
目录结构
----
+ public
color.less `动态主题色less文件`
favicon.ico `系统图标`
index.ejs `网页入口模板`
+ src
+ assets `资源存放`
+ components `组件存放`
- Basic `系统基础组件`
index.ts `抛出所有组件`
+ config `系统配置`
cryptoKey.ts `本地记住密码加密的key(基于crypto-js的aes加密)`
default.homeKey.ts `项目中默认的首页id`
menu.ts `菜单的动态生成方法(基于routes.ts)`
routes.ts `路由配置`
routesInfo.ts `生成路由信息`
systemLocalStoreKey.ts `系统基础组件需要缓存的数据key`
themeColor.ts `主题色列表`
+ layouts `布局`
- BasicLayout `系统基础布局(成功登录后)`
- RouterLayout `系统多级路由容器`
index.ts `抛出所有layouts`
+ locale `语言国际化配置`
- en-US `english`
- zh-CN `简体中文`
- zh-TW `繁体中文`
default.ts `初始化默认语言`
index.ts `抛出所有语言`
+ localStore `路由页面数据持久化`
+ store `vuex状态集中管理`
- models `分模块管理状态`
actions.ts `根级action`
clearStore.ts `初始化当前状态`
extendsLocalStore.ts `路由数据持久化时,继承本地数据的方法`
index.ts `生成状态库`
muations `根级mutaion`
types.ts `该目录下用到的类型声明约定`
+ style `样式`
defaultThemeColor.ts `系统初始化时默认颜色`
index.css `全局样式文件`
reset.css `重置样式`
theme.less `全局less变量定义`
+ utils `工具方法`
appendLessFile.ts `添加主题色编译的color.less`
createGuid.ts `创建不重复的guid`
createThemeColorCss.ts `热编译主题色css`
crypto.ts `本地记住密码加解密方法`
enableRouterLocalStore.ts `开启路由数据持久化`
event.ts `处理事件绑定`
getClientHW.ts `获取客户端宽度与高度`
getDeviceType.ts `获取客户端类型`
validateLogin.ts `验证用户是否登录`
+ views `视图容器`
- Login `登录界面`
App.tsx `vue渲染根节点`
main.ts `主入口`
router.ts `路由生成`
shims-other.d.ts `项目中需要添加的typescript配置额外声明`
shims-tsx.d.ts `tsx类型的全局声明`
shims-vue.d.ts `vue相关类型声明`
+ tests `单元测试`
.eslintrc.js `eslint配置文件`
babel.config.js `babel配置文件`
tsconfig.json `typescript配置`
vue.config.js `vue-cli配置重写`
约束
---
* 路由级组件容器置于`src/views`目录下
* 可复用功能性组件置于`src/components`目录下
* 系统配置文件置于`src/config`目录下
* 工具方法置放于`src/utils`目录下
系统配置
---
|
Subsets and Splits
PyTorch Neural Network Imports
This query filters for code examples containing a specific PyTorch import pattern, which is useful for finding code snippets that use PyTorch's neural network module but doesn't provide deeper analytical insights about the dataset.
HTML Files in Train Set
Retrieves all records from the dataset where the file path ends with .html or .htm, providing a basic filter for HTML files.
SQL Console for nick007x/github-code-2025
Retrieves 200 file paths that end with '.html' or '.htm', providing a basic overview of HTML files in the dataset.
Top HTML Files
The query retrieves a sample of HTML file paths, providing basic filtering but limited analytical value.
CSharp Repositories Excluding Unity
Retrieves all records for repositories that contain C# files but are not related to Unity, providing a basic filter of the dataset.
C# File Count per Repository
Counts the total number of C# files across distinct repositories, providing a basic measure of C# file presence.
SQL Console for nick007x/github-code-2025
Lists unique repository IDs containing C# files, providing basic filtering to understand which repositories have C# code.
Select Groovy Files: Train Set
Retrieves the first 1000 entries from the 'train' dataset where the file path ends with '.groovy', providing a basic sample of Groovy files.
GitHub Repos with WiFiClientSecure
Finds specific file paths in repositories that contain particular code snippets related to WiFiClientSecure and ChatGPT, providing basic filtering of relevant files.