File size: 6,541 Bytes
b6cd2aa
7ed539d
af22b4b
 
90475ed
78baf01
75a98c8
af22b4b
 
1ef33a6
 
 
 
b6cd2aa
af22b4b
 
 
 
 
2ab6cd2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b6cd2aa
2ab6cd2
 
dfd834e
2ab6cd2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b6cd2aa
85d3c28
2ab6cd2
 
 
 
 
 
 
 
 
 
85d3c28
 
 
2ab6cd2
 
 
 
 
 
 
 
 
 
 
 
 
 
85d3c28
af22b4b
 
 
 
 
1ef33a6
 
 
 
 
 
af22b4b
75a98c8
af22b4b
 
75a98c8
af22b4b
75a98c8
1ef33a6
 
 
 
af22b4b
 
75a98c8
 
af22b4b
b6cd2aa
af22b4b
 
 
 
 
 
1ef33a6
af22b4b
 
 
 
 
 
8e6dd44
af22b4b
85d3c28
af22b4b
 
b6cd2aa
7ed539d
af22b4b
 
85d3c28
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
import re
import yaml
import aiohttp
import asyncio
import datetime
import sys
import traceback
from aiohttp import web
from urllib.parse import parse_qs
from cachetools import TTLCache

# 创建一个TTL缓存,最多存储1000个项目,每个项目的有效期为30分钟
cache = TTLCache(maxsize=1000, ttl=1800)

async def fetch_url(url, session):
    async with session.get(url) as response:
        return await response.text()

async def extract_and_transform_proxies(input_text):
    try:
        data = yaml.safe_load(input_text)
        if isinstance(data, dict) and 'proxies' in data:
            proxies_list = data['proxies']
        elif isinstance(data, list):
            proxies_list = data
        else:
            proxies_match = re.search(r'proxies:\s*\n((?:[-\s]*{.*\n?)*)', input_text, re.MULTILINE)
            if proxies_match:
                proxies_text = proxies_match.group(1)
                proxies_list = yaml.safe_load(proxies_text)
            else:
                return "未找到有效的代理配置"
    except yaml.YAMLError:
        return "YAML解析错误"

    if not proxies_list:
        return "未找到有效的代理配置"

    transformed_proxies = []

    for proxy in proxies_list:
        if proxy.get('type') == 'ss':
            name = proxy.get('name', '').strip()
            server = proxy.get('server', '').strip()
            port = str(proxy.get('port', '')).strip()
            
            ss_parts = [f"{name} = ss, {server}, {port}"]
            
            if 'cipher' in proxy:
                ss_parts.append(f"encrypt-method={proxy['cipher'].strip()}")
            if 'password' in proxy:
                ss_parts.append(f"password={proxy['password'].strip()}")
            if 'udp' in proxy:
                ss_parts.append(f"udp-relay={'true' if proxy['udp'] in [True, 'true', 'True'] else 'false'}")

            transformed = ", ".join(ss_parts)
            transformed_proxies.append(transformed)

        elif proxy.get('type') == 'trojan':
            name = proxy.get('name', '').strip()
            server = proxy.get('server', '').strip()
            port = str(proxy.get('port', '')).strip()
            
            trojan_parts = [f"{name} = trojan, {server}, {port}"]
            
            if 'password' in proxy:
                trojan_parts.append(f"password={proxy['password'].strip()}")
            if 'sni' in proxy:
                trojan_parts.append(f"sni={proxy['sni'].strip()}")
            if 'skip-cert-verify' in proxy:
                trojan_parts.append(f"skip-cert-verify={str(proxy['skip-cert-verify']).lower()}")
            if 'udp' in proxy:
                trojan_parts.append(f"udp={'true' if proxy['udp'] in [True, 'true', 'True'] else 'false'}")

            transformed = ", ".join(trojan_parts)
            transformed_proxies.append(transformed)

    return "\n".join(transformed_proxies) if transformed_proxies else "未找到有效的SS或Trojan代理配置"

async def log_request(request, response):
    timestamp = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')
    client_ip = request.remote
    request_line = f"{request.method} {request.path}"
    if request.query_string:
        request_line += f"?{request.query_string}"
    status_code = response.status
    content_length = response.content_length

    log_message = f"{timestamp} - {client_ip} - \"{request_line}\" {status_code} {content_length}"
    print(log_message, flush=True)

@web.middleware
async def logging_middleware(request, handler):
    start_time = datetime.datetime.now()
    try:
        response = await handler(request)
        await log_request(request, response)
        end_time = datetime.datetime.now()
        print(f"Request processing time: {end_time - start_time}", flush=True)
        return response
    except Exception as e:
        end_time = datetime.datetime.now()
        print(f"Error occurred: {str(e)}", flush=True)
        print(f"Request processing time: {end_time - start_time}", flush=True)
        print("Traceback:", flush=True)
        traceback.print_exc()
        return web.Response(text=f"Internal Server Error: {str(e)}", status=500)

async def handle_request(request):
    if request.path == '/':
        query_params = parse_qs(request.query_string)
        if 'url' in query_params:
            url = query_params['url'][0]
            force_refresh = 'nocache' in query_params
            
            if not force_refresh and url in cache:
                print(f"Cache hit for URL: {url}", flush=True)
                return web.Response(text=cache[url], content_type='text/plain')
            
            try:
                print(f"Fetching URL: {url}", flush=True)
                async with aiohttp.ClientSession() as session:
                    input_text = await fetch_url(url, session)
                print(f"URL content length: {len(input_text)}", flush=True)
                result = await extract_and_transform_proxies(input_text)
                print(f"Transformed result length: {len(result)}", flush=True)
                
                # 将结果存入缓存
                cache[url] = result
                
                return web.Response(text=result, content_type='text/plain')
            except Exception as e:
                print(f"Error processing request: {str(e)}", flush=True)
                traceback.print_exc()
                return web.Response(text=f"Error: {str(e)}", status=500)
        else:
            usage_guide = """
            <html>
            <body>
            <h1>代理配置转换工具</h1>
            <p>使用方法:在URL参数中提供包含代理配置的网址。</p>
            <p>示例:<code>http://localhost:8080/?url=https://example.com/path-to-proxy-config</code></p>
            <p>强制刷新缓存:<code>http://localhost:8080/?url=https://example.com/path-to-proxy-config&nocache</code></p>
            </body>
            </html>
            """
            return web.Response(text=usage_guide, content_type='text/html')
    else:
        return web.Response(text="Not Found", status=404)

async def init_app():
    app = web.Application(middlewares=[logging_middleware])
    app.router.add_get('/', handle_request)
    return app

if __name__ == "__main__":
    print(f"===== Application Startup at {datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')} =====")
    print("Server running on port 8080")
    web.run_app(init_app(), port=8080, print=lambda _: None)  # Disable default startup message