Compare commits
2 Commits
005de98ecc
...
master
| Author | SHA1 | Date | |
|---|---|---|---|
| 1c7a6a1f5c | |||
| f0114ef33c |
5
.gitignore
vendored
5
.gitignore
vendored
@ -157,10 +157,6 @@ dmypy.json
|
|||||||
# Cython debug symbols
|
# Cython debug symbols
|
||||||
cython_debug/
|
cython_debug/
|
||||||
|
|
||||||
# migration files
|
|
||||||
rog/migrations/
|
|
||||||
|
|
||||||
|
|
||||||
# PyCharm
|
# PyCharm
|
||||||
# JetBrains specific template is maintainted in a separate JetBrains.gitignore that can
|
# JetBrains specific template is maintainted in a separate JetBrains.gitignore that can
|
||||||
# be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore
|
# be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore
|
||||||
@ -169,4 +165,3 @@ rog/migrations/
|
|||||||
#.idea/
|
#.idea/
|
||||||
|
|
||||||
# End of https://www.toptal.com/developers/gitignore/api/django
|
# End of https://www.toptal.com/developers/gitignore/api/django
|
||||||
.DS_Store
|
|
||||||
|
|||||||
@ -1,9 +1,11 @@
|
|||||||
# FROM python:3.9.9-slim-buster
|
# FROM python:3.9.9-slim-buster
|
||||||
FROM osgeo/gdal:ubuntu-small-3.4.0
|
FROM osgeo/gdal:ubuntu-small-3.4.0
|
||||||
|
# Akira
|
||||||
|
FROM python:3.10
|
||||||
|
FROM ubuntu:latest
|
||||||
|
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
|
|
||||||
|
|
||||||
LABEL maintainer="nouffer@gmail.com"
|
LABEL maintainer="nouffer@gmail.com"
|
||||||
LABEL description="Development image for the Rogaining JP"
|
LABEL description="Development image for the Rogaining JP"
|
||||||
|
|
||||||
@ -15,6 +17,13 @@ ARG TZ Asia/Tokyo \
|
|||||||
|
|
||||||
RUN apt-get update -y
|
RUN apt-get update -y
|
||||||
|
|
||||||
|
# 必要なライブラリのインストール by akira
|
||||||
|
RUN apt-get update && \
|
||||||
|
apt-get install -y software-properties-common && \
|
||||||
|
add-apt-repository ppa:ubuntugis/ppa && \
|
||||||
|
apt-get update && \
|
||||||
|
apt-get install -y gdal-bin libgdal-dev python3-gdal
|
||||||
|
|
||||||
# Install GDAL dependencies
|
# Install GDAL dependencies
|
||||||
RUN apt-get install -y libgdal-dev g++ --no-install-recommends && \
|
RUN apt-get install -y libgdal-dev g++ --no-install-recommends && \
|
||||||
apt-get clean -y
|
apt-get clean -y
|
||||||
@ -24,7 +33,7 @@ ENV CPLUS_INCLUDE_PATH=/usr/include/gdal
|
|||||||
ENV C_INCLUDE_PATH=/usr/include/gdal
|
ENV C_INCLUDE_PATH=/usr/include/gdal
|
||||||
|
|
||||||
RUN apt-get update \
|
RUN apt-get update \
|
||||||
&& apt-get -y install netcat gcc postgresql \
|
&& apt-get -y install netcat-openbsd gcc postgresql \
|
||||||
&& apt-get clean
|
&& apt-get clean
|
||||||
|
|
||||||
RUN apt-get update \
|
RUN apt-get update \
|
||||||
@ -39,63 +48,12 @@ RUN apt-get install -y python3
|
|||||||
RUN apt-get update && apt-get install -y \
|
RUN apt-get update && apt-get install -y \
|
||||||
python3-pip
|
python3-pip
|
||||||
|
|
||||||
# ベースイメージの更新とパッケージのインストール
|
|
||||||
RUN apt-get update && \
|
|
||||||
apt-get install -y \
|
|
||||||
libreoffice \
|
|
||||||
libreoffice-calc \
|
|
||||||
libreoffice-writer \
|
|
||||||
libreoffice-java-common \
|
|
||||||
fonts-ipafont \
|
|
||||||
fonts-ipafont-gothic \
|
|
||||||
fonts-ipafont-mincho \
|
|
||||||
language-pack-ja \
|
|
||||||
fontconfig \
|
|
||||||
locales \
|
|
||||||
python3-uno # LibreOffice Python バインディング
|
|
||||||
|
|
||||||
|
|
||||||
# 日本語ロケールの設定
|
|
||||||
RUN locale-gen ja_JP.UTF-8
|
|
||||||
ENV LANG=ja_JP.UTF-8
|
|
||||||
ENV LC_ALL=ja_JP.UTF-8
|
|
||||||
ENV LANGUAGE=ja_JP:ja
|
|
||||||
|
|
||||||
# フォント設定ファイルをコピー
|
|
||||||
COPY config/fonts.conf /etc/fonts/local.conf
|
|
||||||
|
|
||||||
# フォントキャッシュの更新
|
|
||||||
RUN fc-cache -f -v
|
|
||||||
|
|
||||||
# LibreOfficeの作業ディレクトリを作成
|
|
||||||
RUN mkdir -p /var/cache/libreoffice && \
|
|
||||||
chmod 777 /var/cache/libreoffice
|
|
||||||
|
|
||||||
# フォント設定の権限を設定
|
|
||||||
RUN chmod 644 /etc/fonts/local.conf
|
|
||||||
|
|
||||||
|
|
||||||
# 作業ディレクトリとパーミッションの設定
|
|
||||||
RUN mkdir -p /app/docbase /tmp/libreoffice && \
|
|
||||||
chmod -R 777 /app/docbase /tmp/libreoffice
|
|
||||||
|
|
||||||
|
|
||||||
RUN pip install --upgrade pip
|
RUN pip install --upgrade pip
|
||||||
|
|
||||||
# Copy the package directory first
|
|
||||||
COPY SumasenLibs/excel_lib /app/SumasenLibs/excel_lib
|
|
||||||
COPY ./docbase /app/docbase
|
|
||||||
|
|
||||||
# Install the package in editable mode
|
|
||||||
RUN pip install -e /app/SumasenLibs/excel_lib
|
|
||||||
|
|
||||||
|
|
||||||
RUN apt-get update
|
RUN apt-get update
|
||||||
|
|
||||||
COPY ./requirements.txt /app/requirements.txt
|
COPY ./requirements.txt /app/requirements.txt
|
||||||
|
|
||||||
RUN pip install boto3==1.26.137
|
|
||||||
|
|
||||||
# Install Gunicorn
|
# Install Gunicorn
|
||||||
RUN pip install gunicorn
|
RUN pip install gunicorn
|
||||||
|
|
||||||
@ -103,15 +61,14 @@ RUN pip install gunicorn
|
|||||||
|
|
||||||
#RUN ["chmod", "+x", "wait-for.sh"]
|
#RUN ["chmod", "+x", "wait-for.sh"]
|
||||||
|
|
||||||
# xlsxwriterを追加
|
RUN pip install -r requirements.txt
|
||||||
RUN pip install -r requirements.txt \
|
|
||||||
&& pip install django-cors-headers \
|
|
||||||
&& pip install xlsxwriter gunicorn
|
|
||||||
|
|
||||||
COPY . /app
|
COPY . /app
|
||||||
|
|
||||||
# Collect static files
|
# Collect static files
|
||||||
RUN python manage.py collectstatic --noinput
|
RUN python3 manage.py collectstatic --noinput
|
||||||
|
|
||||||
# Use Gunicorn as the entrypoint
|
# Use Gunicorn as the entrypoint
|
||||||
CMD ["gunicorn", "config.wsgi:application", "--bind", "0.0.0.0:8000"]
|
CMD ["gunicorn", "config.wsgi:application", "--bind", "0.0.0.0:8000"]
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@ -1,35 +0,0 @@
|
|||||||
FROM nginx:alpine
|
|
||||||
|
|
||||||
# Create necessary directories and set permissions
|
|
||||||
RUN mkdir -p /usr/share/nginx/html \
|
|
||||||
&& mkdir -p /var/log/nginx \
|
|
||||||
&& mkdir -p /var/cache/nginx \
|
|
||||||
&& chown -R nginx:nginx /usr/share/nginx/html \
|
|
||||||
&& chown -R nginx:nginx /var/log/nginx \
|
|
||||||
&& chown -R nginx:nginx /var/cache/nginx \
|
|
||||||
&& chmod -R 755 /usr/share/nginx/html
|
|
||||||
|
|
||||||
# Copy files - notice the change in the source path
|
|
||||||
COPY supervisor/html/* /usr/share/nginx/html/
|
|
||||||
COPY supervisor/nginx/default.conf /etc/nginx/conf.d/default.conf
|
|
||||||
|
|
||||||
# メディアディレクトリを作成
|
|
||||||
RUN mkdir -p /app/media && chmod 755 /app/media
|
|
||||||
|
|
||||||
# 静的ファイルをコピー
|
|
||||||
#COPY ./static /usr/share/nginx/html/static
|
|
||||||
|
|
||||||
# 権限の設定
|
|
||||||
RUN chown -R nginx:nginx /app/media
|
|
||||||
|
|
||||||
# Set final permissions
|
|
||||||
RUN chown -R nginx:nginx /usr/share/nginx/html \
|
|
||||||
&& chmod -R 755 /usr/share/nginx/html \
|
|
||||||
&& touch /var/log/nginx/access.log \
|
|
||||||
&& touch /var/log/nginx/error.log \
|
|
||||||
&& chown -R nginx:nginx /var/log/nginx \
|
|
||||||
&& chown -R nginx:nginx /etc/nginx/conf.d
|
|
||||||
|
|
||||||
EXPOSE 80
|
|
||||||
|
|
||||||
CMD ["nginx", "-g", "daemon off;"]
|
|
||||||
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
6
README.jpn
Normal file
6
README.jpn
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
デプロイ:
|
||||||
|
you can just run
|
||||||
|
docker-compose up -d
|
||||||
|
will deploy it
|
||||||
|
|
||||||
|
|
||||||
Binary file not shown.
@ -1,19 +0,0 @@
|
|||||||
# SumasenExcel Library
|
|
||||||
|
|
||||||
Excel操作のためのシンプルなPythonライブラリです。
|
|
||||||
|
|
||||||
## インストール方法
|
|
||||||
|
|
||||||
```bash
|
|
||||||
pip install -e .
|
|
||||||
|
|
||||||
## 使用方法
|
|
||||||
from sumaexcel import SumasenExcel
|
|
||||||
|
|
||||||
excel = SumasenExcel("path/to/file.xlsx")
|
|
||||||
data = excel.read_excel()
|
|
||||||
|
|
||||||
## ライセンス
|
|
||||||
|
|
||||||
MIT License
|
|
||||||
|
|
||||||
@ -1,20 +0,0 @@
|
|||||||
version: '3.8'
|
|
||||||
|
|
||||||
services:
|
|
||||||
python:
|
|
||||||
build:
|
|
||||||
context: ..
|
|
||||||
dockerfile: docker/python/Dockerfile
|
|
||||||
volumes:
|
|
||||||
- ..:/app
|
|
||||||
environment:
|
|
||||||
- PYTHONPATH=/app
|
|
||||||
- POSTGRES_DB=rogdb
|
|
||||||
- POSTGRES_USER=admin
|
|
||||||
- POSTGRES_PASSWORD=admin123456
|
|
||||||
- POSTGRES_HOST=localhost
|
|
||||||
- POSTGRES_PORT=5432
|
|
||||||
network_mode: "host"
|
|
||||||
tty: true
|
|
||||||
container_name: python_container # コンテナ名を明示的に指定
|
|
||||||
|
|
||||||
@ -1,26 +0,0 @@
|
|||||||
FROM python:3.9-slim
|
|
||||||
|
|
||||||
WORKDIR /app
|
|
||||||
|
|
||||||
# GPGキーの更新とパッケージのインストール
|
|
||||||
RUN apt-get update --allow-insecure-repositories && \
|
|
||||||
apt-get install -y --allow-unauthenticated python3-dev libpq-dev postgresql-client && \
|
|
||||||
rm -rf /var/lib/apt/lists/*
|
|
||||||
|
|
||||||
# Pythonパッケージのインストール
|
|
||||||
COPY requirements.txt .
|
|
||||||
COPY setup.py .
|
|
||||||
COPY README.md .
|
|
||||||
COPY . .
|
|
||||||
|
|
||||||
RUN pip install --no-cache-dir -r requirements.txt
|
|
||||||
|
|
||||||
# 開発用パッケージのインストール
|
|
||||||
RUN pip install --no-cache-dir --upgrade pip \
|
|
||||||
pytest \
|
|
||||||
pytest-cov \
|
|
||||||
flake8
|
|
||||||
|
|
||||||
# パッケージのインストール
|
|
||||||
RUN pip install -e .
|
|
||||||
|
|
||||||
@ -1,6 +0,0 @@
|
|||||||
openpyxl>=3.0.0
|
|
||||||
pandas>=1.0.0
|
|
||||||
pillow>=8.0.0
|
|
||||||
configparser>=5.0.0
|
|
||||||
psycopg2-binary==2.9.9
|
|
||||||
requests
|
|
||||||
@ -1,25 +0,0 @@
|
|||||||
# setup.py
|
|
||||||
from setuptools import setup, find_packages
|
|
||||||
|
|
||||||
setup(
|
|
||||||
name="sumaexcel",
|
|
||||||
version="0.1.0",
|
|
||||||
packages=find_packages(),
|
|
||||||
install_requires=[
|
|
||||||
"openpyxl>=3.0.0",
|
|
||||||
"pandas>=1.0.0"
|
|
||||||
],
|
|
||||||
author="Akira Miyata",
|
|
||||||
author_email="akira.miyata@sumasen.net",
|
|
||||||
description="Excel handling library",
|
|
||||||
long_description=open("README.md").read(),
|
|
||||||
long_description_content_type="text/markdown",
|
|
||||||
url="https://github.com/akiramiyata/sumaexcel",
|
|
||||||
classifiers=[
|
|
||||||
"Programming Language :: Python :: 3",
|
|
||||||
"License :: OSI Approved :: MIT License",
|
|
||||||
"Operating System :: OS Independent",
|
|
||||||
],
|
|
||||||
python_requires=">=3.6",
|
|
||||||
)
|
|
||||||
|
|
||||||
@ -1,4 +0,0 @@
|
|||||||
from .sumaexcel import SumasenExcel
|
|
||||||
|
|
||||||
__version__ = "0.1.0"
|
|
||||||
__all__ = ["SumasenExcel"]
|
|
||||||
@ -1,102 +0,0 @@
|
|||||||
# sumaexcel/conditional.py
|
|
||||||
from typing import Dict, Any, List, Union
|
|
||||||
from openpyxl.formatting.rule import Rule, ColorScaleRule, DataBarRule, IconSetRule
|
|
||||||
from openpyxl.styles import PatternFill, Font, Border, Side
|
|
||||||
from openpyxl.worksheet.worksheet import Worksheet
|
|
||||||
|
|
||||||
class ConditionalFormatManager:
|
|
||||||
"""Handle conditional formatting in Excel"""
|
|
||||||
|
|
||||||
def __init__(self, worksheet: Worksheet):
|
|
||||||
self.worksheet = worksheet
|
|
||||||
|
|
||||||
def add_color_scale(
|
|
||||||
self,
|
|
||||||
cell_range: str,
|
|
||||||
min_color: str = "00FF0000", # Red
|
|
||||||
mid_color: str = "00FFFF00", # Yellow
|
|
||||||
max_color: str = "0000FF00" # Green
|
|
||||||
) -> None:
|
|
||||||
"""Add color scale conditional formatting"""
|
|
||||||
rule = ColorScaleRule(
|
|
||||||
start_type='min',
|
|
||||||
start_color=min_color,
|
|
||||||
mid_type='percentile',
|
|
||||||
mid_value=50,
|
|
||||||
mid_color=mid_color,
|
|
||||||
end_type='max',
|
|
||||||
end_color=max_color
|
|
||||||
)
|
|
||||||
self.worksheet.conditional_formatting.add(cell_range, rule)
|
|
||||||
|
|
||||||
def add_data_bar(
|
|
||||||
self,
|
|
||||||
cell_range: str,
|
|
||||||
color: str = "000000FF", # Blue
|
|
||||||
show_value: bool = True
|
|
||||||
) -> None:
|
|
||||||
"""Add data bar conditional formatting"""
|
|
||||||
rule = DataBarRule(
|
|
||||||
start_type='min',
|
|
||||||
end_type='max',
|
|
||||||
color=color,
|
|
||||||
showValue=show_value
|
|
||||||
)
|
|
||||||
self.worksheet.conditional_formatting.add(cell_range, rule)
|
|
||||||
|
|
||||||
def add_icon_set(
|
|
||||||
self,
|
|
||||||
cell_range: str,
|
|
||||||
icon_style: str = '3Arrows', # '3Arrows', '3TrafficLights', '3Signs'
|
|
||||||
reverse_icons: bool = False
|
|
||||||
) -> None:
|
|
||||||
"""Add icon set conditional formatting"""
|
|
||||||
rule = IconSetRule(
|
|
||||||
icon_style=icon_style,
|
|
||||||
type='percent',
|
|
||||||
values=[0, 33, 67],
|
|
||||||
reverse_icons=reverse_icons
|
|
||||||
)
|
|
||||||
self.worksheet.conditional_formatting.add(cell_range, rule)
|
|
||||||
|
|
||||||
def add_custom_rule(
|
|
||||||
self,
|
|
||||||
cell_range: str,
|
|
||||||
rule_type: str,
|
|
||||||
formula: str,
|
|
||||||
fill_color: str = None,
|
|
||||||
font_color: str = None,
|
|
||||||
bold: bool = None,
|
|
||||||
border_style: str = None,
|
|
||||||
border_color: str = None
|
|
||||||
) -> None:
|
|
||||||
"""Add custom conditional formatting rule"""
|
|
||||||
dxf = {}
|
|
||||||
if fill_color:
|
|
||||||
dxf['fill'] = PatternFill(start_color=fill_color, end_color=fill_color)
|
|
||||||
if font_color or bold is not None:
|
|
||||||
dxf['font'] = Font(color=font_color, bold=bold)
|
|
||||||
if border_style and border_color:
|
|
||||||
side = Side(style=border_style, color=border_color)
|
|
||||||
dxf['border'] = Border(left=side, right=side, top=side, bottom=side)
|
|
||||||
|
|
||||||
rule = Rule(type=rule_type, formula=[formula], dxf=dxf)
|
|
||||||
self.worksheet.conditional_formatting.add(cell_range, rule)
|
|
||||||
|
|
||||||
def copy_conditional_format(
|
|
||||||
self,
|
|
||||||
source_range: str,
|
|
||||||
target_range: str
|
|
||||||
) -> None:
|
|
||||||
"""Copy conditional formatting from one range to another"""
|
|
||||||
source_rules = self.worksheet.conditional_formatting.get(source_range)
|
|
||||||
if source_rules:
|
|
||||||
for rule in source_rules:
|
|
||||||
self.worksheet.conditional_formatting.add(target_range, rule)
|
|
||||||
|
|
||||||
def clear_conditional_format(
|
|
||||||
self,
|
|
||||||
cell_range: str
|
|
||||||
) -> None:
|
|
||||||
"""Clear conditional formatting from specified range"""
|
|
||||||
self.worksheet.conditional_formatting.delete(cell_range)
|
|
||||||
@ -1,166 +0,0 @@
|
|||||||
# config_handler.py
|
|
||||||
#
|
|
||||||
import configparser
|
|
||||||
import os
|
|
||||||
from typing import Any, Dict, Optional
|
|
||||||
|
|
||||||
import configparser
|
|
||||||
import os
|
|
||||||
import re
|
|
||||||
from typing import Any, Dict, Optional
|
|
||||||
|
|
||||||
class ConfigHandler:
|
|
||||||
"""変数置換機能付きの設定ファイル管理クラス"""
|
|
||||||
|
|
||||||
def __init__(self, ini_file_path: str, variables: Dict[str, str] = None):
|
|
||||||
"""
|
|
||||||
Args:
|
|
||||||
ini_file_path (str): INIファイルのパス
|
|
||||||
variables (Dict[str, str], optional): 置換用の変数辞書
|
|
||||||
"""
|
|
||||||
self.ini_file_path = ini_file_path
|
|
||||||
self.variables = variables or {}
|
|
||||||
self.config = configparser.ConfigParser()
|
|
||||||
self.load_config()
|
|
||||||
|
|
||||||
def _substitute_variables(self, text: str) -> str:
|
|
||||||
"""
|
|
||||||
テキスト内の変数を置換する
|
|
||||||
|
|
||||||
Args:
|
|
||||||
text (str): 置換対象のテキスト
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
str: 置換後のテキスト
|
|
||||||
"""
|
|
||||||
# ${var}形式の変数を置換
|
|
||||||
pattern1 = r'\${([^}]+)}'
|
|
||||||
# [var]形式の変数を置換
|
|
||||||
pattern2 = r'\[([^\]]+)\]'
|
|
||||||
|
|
||||||
def replace_var(match):
|
|
||||||
var_name = match.group(1)
|
|
||||||
return self.variables.get(var_name, match.group(0))
|
|
||||||
|
|
||||||
# 両方のパターンで置換を実行
|
|
||||||
text = re.sub(pattern1, replace_var, text)
|
|
||||||
text = re.sub(pattern2, replace_var, text)
|
|
||||||
|
|
||||||
return text
|
|
||||||
|
|
||||||
def load_config(self) -> None:
|
|
||||||
"""設定ファイルを読み込み、変数を置換する"""
|
|
||||||
if not os.path.exists(self.ini_file_path):
|
|
||||||
raise FileNotFoundError(f"設定ファイルが見つかりません: {self.ini_file_path}")
|
|
||||||
|
|
||||||
# まず生のテキストとして読み込む
|
|
||||||
with open(self.ini_file_path, 'r', encoding='utf-8') as f:
|
|
||||||
content = f.read()
|
|
||||||
|
|
||||||
# 変数を置換
|
|
||||||
substituted_content = self._substitute_variables(content)
|
|
||||||
|
|
||||||
# 置換済みの内容を StringIO 経由で configparser に読み込ませる
|
|
||||||
from io import StringIO
|
|
||||||
self.config.read_file(StringIO(substituted_content))
|
|
||||||
|
|
||||||
def get_value(self, section: str, key: str, default: Any = None) -> Optional[str]:
|
|
||||||
"""
|
|
||||||
指定されたセクションのキーの値を取得する
|
|
||||||
|
|
||||||
Args:
|
|
||||||
section (str): セクション名
|
|
||||||
key (str): キー名
|
|
||||||
default (Any): デフォルト値(オプション)
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Optional[str]: 設定値。存在しない場合はデフォルト値
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
return self.config[section][key]
|
|
||||||
except KeyError:
|
|
||||||
return default
|
|
||||||
|
|
||||||
def get_section(self, section: str) -> Dict[str, str]:
|
|
||||||
"""
|
|
||||||
指定されたセクションの全ての設定を取得する
|
|
||||||
|
|
||||||
Args:
|
|
||||||
section (str): セクション名
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Dict[str, str]: セクションの設定をディクショナリで返す
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
return dict(self.config[section])
|
|
||||||
except KeyError:
|
|
||||||
return {}
|
|
||||||
|
|
||||||
def get_all_sections(self) -> Dict[str, Dict[str, str]]:
|
|
||||||
"""
|
|
||||||
全てのセクションの設定を取得する
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Dict[str, Dict[str, str]]: 全セクションの設定をネストされたディクショナリで返す
|
|
||||||
"""
|
|
||||||
return {section: dict(self.config[section]) for section in self.config.sections()}
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# 使用例
|
|
||||||
if __name__ == "__main__":
|
|
||||||
# サンプルのINIファイル作成
|
|
||||||
sample_ini = """
|
|
||||||
[Database]
|
|
||||||
host = localhost
|
|
||||||
port = 5432
|
|
||||||
database = mydb
|
|
||||||
user = admin
|
|
||||||
password = secret
|
|
||||||
|
|
||||||
[Application]
|
|
||||||
debug = true
|
|
||||||
log_level = INFO
|
|
||||||
max_connections = 100
|
|
||||||
|
|
||||||
[Paths]
|
|
||||||
data_dir = /var/data
|
|
||||||
log_file = /var/log/app.log
|
|
||||||
"""
|
|
||||||
|
|
||||||
# サンプルINIファイルを作成
|
|
||||||
with open('config.ini', 'w', encoding='utf-8') as f:
|
|
||||||
f.write(sample_ini)
|
|
||||||
|
|
||||||
# 設定を読み込んで使用
|
|
||||||
config = ConfigHandler('config.ini')
|
|
||||||
|
|
||||||
# 特定の値を取得
|
|
||||||
db_host = config.get_value('Database', 'host')
|
|
||||||
db_port = config.get_value('Database', 'port')
|
|
||||||
print(f"Database connection: {db_host}:{db_port}")
|
|
||||||
|
|
||||||
# セクション全体を取得
|
|
||||||
db_config = config.get_section('Database')
|
|
||||||
print("Database configuration:", db_config)
|
|
||||||
|
|
||||||
# 全ての設定を取得
|
|
||||||
all_config = config.get_all_sections()
|
|
||||||
print("All configurations:", all_config)
|
|
||||||
|
|
||||||
|
|
||||||
# サンプル:
|
|
||||||
# # 設定ファイルから値を取得
|
|
||||||
# config = ConfigHandler('config.ini')
|
|
||||||
#
|
|
||||||
# # データベース設定を取得
|
|
||||||
# db_host = config.get_value('Database', 'host')
|
|
||||||
# db_port = config.get_value('Database', 'port')
|
|
||||||
# db_name = config.get_value('Database', 'database')
|
|
||||||
#
|
|
||||||
# # アプリケーション設定を取得
|
|
||||||
# debug_mode = config.get_value('Application', 'debug')
|
|
||||||
# log_level = config.get_value('Application', 'log_level')
|
|
||||||
#
|
|
||||||
@ -1,77 +0,0 @@
|
|||||||
# sumaexcel/image.py
|
|
||||||
from typing import Optional, Tuple, Union
|
|
||||||
from pathlib import Path
|
|
||||||
import os
|
|
||||||
from PIL import Image
|
|
||||||
from openpyxl.drawing.image import Image as XLImage
|
|
||||||
from openpyxl.worksheet.worksheet import Worksheet
|
|
||||||
|
|
||||||
class ImageManager:
|
|
||||||
"""Handle image operations in Excel"""
|
|
||||||
|
|
||||||
def __init__(self, worksheet: Worksheet):
|
|
||||||
self.worksheet = worksheet
|
|
||||||
self.temp_dir = Path("/tmp/sumaexcel_images")
|
|
||||||
self.temp_dir.mkdir(parents=True, exist_ok=True)
|
|
||||||
|
|
||||||
def add_image(
|
|
||||||
self,
|
|
||||||
image_path: Union[str, Path],
|
|
||||||
cell_coordinates: Tuple[int, int],
|
|
||||||
size: Optional[Tuple[int, int]] = None,
|
|
||||||
keep_aspect_ratio: bool = True,
|
|
||||||
anchor_type: str = 'absolute'
|
|
||||||
) -> None:
|
|
||||||
"""Add image to worksheet at specified position"""
|
|
||||||
# Convert path to Path object
|
|
||||||
image_path = Path(image_path)
|
|
||||||
|
|
||||||
# Open and process image
|
|
||||||
with Image.open(image_path) as img:
|
|
||||||
# Get original size
|
|
||||||
orig_width, orig_height = img.size
|
|
||||||
|
|
||||||
# Calculate new size if specified
|
|
||||||
if size:
|
|
||||||
target_width, target_height = size
|
|
||||||
if keep_aspect_ratio:
|
|
||||||
ratio = min(target_width/orig_width, target_height/orig_height)
|
|
||||||
target_width = int(orig_width * ratio)
|
|
||||||
target_height = int(orig_height * ratio)
|
|
||||||
|
|
||||||
# Resize image
|
|
||||||
img = img.resize((target_width, target_height), Image.LANCZOS)
|
|
||||||
|
|
||||||
# Save temporary resized image
|
|
||||||
temp_path = self.temp_dir / f"temp_{image_path.name}"
|
|
||||||
img.save(temp_path)
|
|
||||||
image_path = temp_path
|
|
||||||
|
|
||||||
# Create Excel image object
|
|
||||||
excel_image = XLImage(str(image_path))
|
|
||||||
|
|
||||||
# Add to worksheet
|
|
||||||
self.worksheet.add_image(excel_image, anchor=f'{cell_coordinates[0]}{cell_coordinates[1]}')
|
|
||||||
|
|
||||||
def add_image_absolute(
|
|
||||||
self,
|
|
||||||
image_path: Union[str, Path],
|
|
||||||
position: Tuple[int, int],
|
|
||||||
size: Optional[Tuple[int, int]] = None
|
|
||||||
) -> None:
|
|
||||||
"""Add image with absolute positioning"""
|
|
||||||
excel_image = XLImage(str(image_path))
|
|
||||||
if size:
|
|
||||||
excel_image.width, excel_image.height = size
|
|
||||||
excel_image.anchor = 'absolute'
|
|
||||||
excel_image.top, excel_image.left = position
|
|
||||||
self.worksheet.add_image(excel_image)
|
|
||||||
|
|
||||||
def cleanup(self) -> None:
|
|
||||||
"""Clean up temporary files"""
|
|
||||||
for file in self.temp_dir.glob("temp_*"):
|
|
||||||
file.unlink()
|
|
||||||
|
|
||||||
def __del__(self):
|
|
||||||
"""Cleanup on object destruction"""
|
|
||||||
self.cleanup()
|
|
||||||
@ -1,96 +0,0 @@
|
|||||||
# sumaexcel/merge.py
|
|
||||||
from typing import List, Tuple, Dict
|
|
||||||
from openpyxl.worksheet.worksheet import Worksheet
|
|
||||||
from openpyxl.worksheet.merge import MergedCellRange
|
|
||||||
|
|
||||||
class MergeManager:
|
|
||||||
"""Handle merge cell operations"""
|
|
||||||
|
|
||||||
def __init__(self, worksheet: Worksheet):
|
|
||||||
self.worksheet = worksheet
|
|
||||||
self._merged_ranges: List[MergedCellRange] = []
|
|
||||||
self._load_merged_ranges()
|
|
||||||
|
|
||||||
def _load_merged_ranges(self) -> None:
|
|
||||||
"""Load existing merged ranges from worksheet"""
|
|
||||||
self._merged_ranges = list(self.worksheet.merged_cells.ranges)
|
|
||||||
|
|
||||||
def merge_cells(
|
|
||||||
self,
|
|
||||||
start_row: int,
|
|
||||||
start_col: int,
|
|
||||||
end_row: int,
|
|
||||||
end_col: int
|
|
||||||
) -> None:
|
|
||||||
"""Merge cells in specified range"""
|
|
||||||
self.worksheet.merge_cells(
|
|
||||||
start_row=start_row,
|
|
||||||
start_column=start_col,
|
|
||||||
end_row=end_row,
|
|
||||||
end_column=end_col
|
|
||||||
)
|
|
||||||
self._load_merged_ranges()
|
|
||||||
|
|
||||||
def unmerge_cells(
|
|
||||||
self,
|
|
||||||
start_row: int,
|
|
||||||
start_col: int,
|
|
||||||
end_row: int,
|
|
||||||
end_col: int
|
|
||||||
) -> None:
|
|
||||||
"""Unmerge cells in specified range"""
|
|
||||||
self.worksheet.unmerge_cells(
|
|
||||||
start_row=start_row,
|
|
||||||
start_column=start_col,
|
|
||||||
end_row=end_row,
|
|
||||||
end_column=end_col
|
|
||||||
)
|
|
||||||
self._load_merged_ranges()
|
|
||||||
|
|
||||||
def copy_merged_cells(
|
|
||||||
self,
|
|
||||||
source_range: Tuple[int, int, int, int],
|
|
||||||
target_start_row: int,
|
|
||||||
target_start_col: int
|
|
||||||
) -> None:
|
|
||||||
"""Copy merged cells from source range to target position"""
|
|
||||||
src_row1, src_col1, src_row2, src_col2 = source_range
|
|
||||||
row_offset = target_start_row - src_row1
|
|
||||||
col_offset = target_start_col - src_col1
|
|
||||||
|
|
||||||
for merged_range in self._merged_ranges:
|
|
||||||
if (src_row1 <= merged_range.min_row <= src_row2 and
|
|
||||||
src_col1 <= merged_range.min_col <= src_col2):
|
|
||||||
new_row1 = merged_range.min_row + row_offset
|
|
||||||
new_col1 = merged_range.min_col + col_offset
|
|
||||||
new_row2 = merged_range.max_row + row_offset
|
|
||||||
new_col2 = merged_range.max_col + col_offset
|
|
||||||
|
|
||||||
self.merge_cells(new_row1, new_col1, new_row2, new_col2)
|
|
||||||
|
|
||||||
def shift_merged_cells(
|
|
||||||
self,
|
|
||||||
start_row: int,
|
|
||||||
rows: int = 0,
|
|
||||||
cols: int = 0
|
|
||||||
) -> None:
|
|
||||||
"""Shift merged cells by specified number of rows and columns"""
|
|
||||||
new_ranges = []
|
|
||||||
for merged_range in self._merged_ranges:
|
|
||||||
if merged_range.min_row >= start_row:
|
|
||||||
new_row1 = merged_range.min_row + rows
|
|
||||||
new_col1 = merged_range.min_col + cols
|
|
||||||
new_row2 = merged_range.max_row + rows
|
|
||||||
new_col2 = merged_range.max_col + cols
|
|
||||||
|
|
||||||
self.worksheet.unmerge_cells(
|
|
||||||
start_row=merged_range.min_row,
|
|
||||||
start_column=merged_range.min_col,
|
|
||||||
end_row=merged_range.max_row,
|
|
||||||
end_column=merged_range.max_col
|
|
||||||
)
|
|
||||||
|
|
||||||
new_ranges.append((new_row1, new_col1, new_row2, new_col2))
|
|
||||||
|
|
||||||
for new_range in new_ranges:
|
|
||||||
self.merge_cells(*new_range)
|
|
||||||
@ -1,148 +0,0 @@
|
|||||||
# sumaexcel/page.py
|
|
||||||
from typing import Optional, Dict, Any, Union
|
|
||||||
from openpyxl.worksheet.worksheet import Worksheet
|
|
||||||
from openpyxl.worksheet.page import PageMargins, PrintPageSetup
|
|
||||||
|
|
||||||
# sumaexcel/page.py (continued)
|
|
||||||
|
|
||||||
class PageManager:
|
|
||||||
"""Handle page setup and header/footer settings"""
|
|
||||||
|
|
||||||
def __init__(self, worksheet: Worksheet):
|
|
||||||
self.worksheet = worksheet
|
|
||||||
|
|
||||||
def set_page_setup(
|
|
||||||
self,
|
|
||||||
orientation: str = 'portrait',
|
|
||||||
paper_size: int = 9, # A4
|
|
||||||
fit_to_height: Optional[int] = None,
|
|
||||||
fit_to_width: Optional[int] = None,
|
|
||||||
scale: Optional[int] = None
|
|
||||||
) -> None:
|
|
||||||
"""Configure page setup
|
|
||||||
|
|
||||||
Args:
|
|
||||||
orientation: 'portrait' or 'landscape'
|
|
||||||
paper_size: paper size (e.g., 9 for A4)
|
|
||||||
fit_to_height: number of pages tall
|
|
||||||
fit_to_width: number of pages wide
|
|
||||||
scale: zoom scale (1-400)
|
|
||||||
"""
|
|
||||||
setup = PrintPageSetup(
|
|
||||||
orientation=orientation,
|
|
||||||
paperSize=paper_size,
|
|
||||||
scale=scale,
|
|
||||||
fitToHeight=fit_to_height,
|
|
||||||
fitToWidth=fit_to_width
|
|
||||||
)
|
|
||||||
self.worksheet.page_setup = setup
|
|
||||||
|
|
||||||
def set_margins(
|
|
||||||
self,
|
|
||||||
left: float = 0.7,
|
|
||||||
right: float = 0.7,
|
|
||||||
top: float = 0.75,
|
|
||||||
bottom: float = 0.75,
|
|
||||||
header: float = 0.3,
|
|
||||||
footer: float = 0.3
|
|
||||||
) -> None:
|
|
||||||
"""Set page margins in inches"""
|
|
||||||
margins = PageMargins(
|
|
||||||
left=left,
|
|
||||||
right=right,
|
|
||||||
top=top,
|
|
||||||
bottom=bottom,
|
|
||||||
header=header,
|
|
||||||
footer=footer
|
|
||||||
)
|
|
||||||
self.worksheet.page_margins = margins
|
|
||||||
|
|
||||||
def set_header_footer(
|
|
||||||
self,
|
|
||||||
odd_header: Optional[str] = None,
|
|
||||||
odd_footer: Optional[str] = None,
|
|
||||||
even_header: Optional[str] = None,
|
|
||||||
even_footer: Optional[str] = None,
|
|
||||||
first_header: Optional[str] = None,
|
|
||||||
first_footer: Optional[str] = None,
|
|
||||||
different_first: bool = False,
|
|
||||||
different_odd_even: bool = False
|
|
||||||
) -> None:
|
|
||||||
"""Set headers and footers
|
|
||||||
|
|
||||||
Format codes:
|
|
||||||
- &P: Page number
|
|
||||||
- &N: Total pages
|
|
||||||
- &D: Date
|
|
||||||
- &T: Time
|
|
||||||
- &[Tab]: Sheet name
|
|
||||||
- &[Path]: File path
|
|
||||||
- &[File]: File name
|
|
||||||
- &[Tab]: Worksheet name
|
|
||||||
"""
|
|
||||||
self.worksheet.oddHeader.left = odd_header or ""
|
|
||||||
self.worksheet.oddFooter.left = odd_footer or ""
|
|
||||||
|
|
||||||
if different_odd_even:
|
|
||||||
self.worksheet.evenHeader.left = even_header or ""
|
|
||||||
self.worksheet.evenFooter.left = even_footer or ""
|
|
||||||
|
|
||||||
if different_first:
|
|
||||||
self.worksheet.firstHeader.left = first_header or ""
|
|
||||||
self.worksheet.firstFooter.left = first_footer or ""
|
|
||||||
|
|
||||||
self.worksheet.differentFirst = different_first
|
|
||||||
self.worksheet.differentOddEven = different_odd_even
|
|
||||||
|
|
||||||
def set_print_area(self, range_string: str) -> None:
|
|
||||||
"""Set print area
|
|
||||||
|
|
||||||
Args:
|
|
||||||
range_string: Cell range in A1 notation (e.g., 'A1:H42')
|
|
||||||
"""
|
|
||||||
self.worksheet.print_area = range_string
|
|
||||||
|
|
||||||
def set_print_title_rows(self, rows: str) -> None:
|
|
||||||
"""Set rows to repeat at top of each page
|
|
||||||
|
|
||||||
Args:
|
|
||||||
rows: Row range (e.g., '1:3')
|
|
||||||
"""
|
|
||||||
self.worksheet.print_title_rows = rows
|
|
||||||
|
|
||||||
def set_print_title_columns(self, cols: str) -> None:
|
|
||||||
"""Set columns to repeat at left of each page
|
|
||||||
|
|
||||||
Args:
|
|
||||||
cols: Column range (e.g., 'A:B')
|
|
||||||
"""
|
|
||||||
self.worksheet.print_title_cols = cols
|
|
||||||
|
|
||||||
def set_print_options(
|
|
||||||
self,
|
|
||||||
grid_lines: bool = False,
|
|
||||||
horizontal_centered: bool = False,
|
|
||||||
vertical_centered: bool = False,
|
|
||||||
headers: bool = False
|
|
||||||
) -> None:
|
|
||||||
"""Set print options"""
|
|
||||||
self.worksheet.print_gridlines = grid_lines
|
|
||||||
self.worksheet.print_options.horizontalCentered = horizontal_centered
|
|
||||||
self.worksheet.print_options.verticalCentered = vertical_centered
|
|
||||||
self.worksheet.print_options.headers = headers
|
|
||||||
|
|
||||||
class PaperSizes:
|
|
||||||
"""Standard paper size constants"""
|
|
||||||
LETTER = 1
|
|
||||||
LETTER_SMALL = 2
|
|
||||||
TABLOID = 3
|
|
||||||
LEDGER = 4
|
|
||||||
LEGAL = 5
|
|
||||||
STATEMENT = 6
|
|
||||||
EXECUTIVE = 7
|
|
||||||
A3 = 8
|
|
||||||
A4 = 9
|
|
||||||
A4_SMALL = 10
|
|
||||||
A5 = 11
|
|
||||||
B4 = 12
|
|
||||||
B5 = 13
|
|
||||||
@ -1,115 +0,0 @@
|
|||||||
# sumaexcel/styles.py
|
|
||||||
from typing import Dict, Any, Optional, Union
|
|
||||||
from openpyxl.styles import (
|
|
||||||
Font, PatternFill, Alignment, Border, Side,
|
|
||||||
NamedStyle, Protection, Color
|
|
||||||
)
|
|
||||||
from openpyxl.styles.differential import DifferentialStyle
|
|
||||||
from openpyxl.formatting.rule import Rule
|
|
||||||
from openpyxl.worksheet.worksheet import Worksheet
|
|
||||||
|
|
||||||
class StyleManager:
|
|
||||||
"""Excel style management class"""
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def create_font(
|
|
||||||
name: str = "Arial",
|
|
||||||
size: int = 11,
|
|
||||||
bold: bool = False,
|
|
||||||
italic: bool = False,
|
|
||||||
color: str = "000000",
|
|
||||||
underline: str = None,
|
|
||||||
strike: bool = False
|
|
||||||
) -> Font:
|
|
||||||
"""Create a Font object with specified parameters"""
|
|
||||||
return Font(
|
|
||||||
name=name,
|
|
||||||
size=size,
|
|
||||||
bold=bold,
|
|
||||||
italic=italic,
|
|
||||||
color=color,
|
|
||||||
underline=underline,
|
|
||||||
strike=strike
|
|
||||||
)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def create_fill(
|
|
||||||
fill_type: str = "solid",
|
|
||||||
start_color: str = "FFFFFF",
|
|
||||||
end_color: str = None
|
|
||||||
) -> PatternFill:
|
|
||||||
"""Create a PatternFill object"""
|
|
||||||
return PatternFill(
|
|
||||||
fill_type=fill_type,
|
|
||||||
start_color=start_color,
|
|
||||||
end_color=end_color or start_color
|
|
||||||
)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def create_border(
|
|
||||||
style: str = "thin",
|
|
||||||
color: str = "000000"
|
|
||||||
) -> Border:
|
|
||||||
"""Create a Border object"""
|
|
||||||
side = Side(style=style, color=color)
|
|
||||||
return Border(
|
|
||||||
left=side,
|
|
||||||
right=side,
|
|
||||||
top=side,
|
|
||||||
bottom=side
|
|
||||||
)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def create_alignment(
|
|
||||||
horizontal: str = "general",
|
|
||||||
vertical: str = "bottom",
|
|
||||||
wrap_text: bool = False,
|
|
||||||
shrink_to_fit: bool = False,
|
|
||||||
indent: int = 0
|
|
||||||
) -> Alignment:
|
|
||||||
"""Create an Alignment object"""
|
|
||||||
return Alignment(
|
|
||||||
horizontal=horizontal,
|
|
||||||
vertical=vertical,
|
|
||||||
wrap_text=wrap_text,
|
|
||||||
shrink_to_fit=shrink_to_fit,
|
|
||||||
indent=indent
|
|
||||||
)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def copy_style(source_cell: Any, target_cell: Any) -> None:
|
|
||||||
"""Copy all style properties from source cell to target cell"""
|
|
||||||
target_cell.font = Font(
|
|
||||||
name=source_cell.font.name,
|
|
||||||
size=source_cell.font.size,
|
|
||||||
bold=source_cell.font.bold,
|
|
||||||
italic=source_cell.font.italic,
|
|
||||||
color=source_cell.font.color,
|
|
||||||
underline=source_cell.font.underline,
|
|
||||||
strike=source_cell.font.strike
|
|
||||||
)
|
|
||||||
|
|
||||||
if source_cell.fill.patternType != None:
|
|
||||||
target_cell.fill = PatternFill(
|
|
||||||
fill_type=source_cell.fill.patternType,
|
|
||||||
start_color=source_cell.fill.start_color.rgb,
|
|
||||||
end_color=source_cell.fill.end_color.rgb
|
|
||||||
)
|
|
||||||
|
|
||||||
target_cell.border = Border(
|
|
||||||
left=source_cell.border.left,
|
|
||||||
right=source_cell.border.right,
|
|
||||||
top=source_cell.border.top,
|
|
||||||
bottom=source_cell.border.bottom
|
|
||||||
)
|
|
||||||
|
|
||||||
target_cell.alignment = Alignment(
|
|
||||||
horizontal=source_cell.alignment.horizontal,
|
|
||||||
vertical=source_cell.alignment.vertical,
|
|
||||||
wrap_text=source_cell.alignment.wrap_text,
|
|
||||||
shrink_to_fit=source_cell.alignment.shrink_to_fit,
|
|
||||||
indent=source_cell.alignment.indent
|
|
||||||
)
|
|
||||||
|
|
||||||
if source_cell.number_format:
|
|
||||||
target_cell.number_format = source_cell.number_format
|
|
||||||
File diff suppressed because it is too large
Load Diff
BIN
SumasenLibs/excel_lib/testdata/certificate_5033.xlsx
vendored
BIN
SumasenLibs/excel_lib/testdata/certificate_5033.xlsx
vendored
Binary file not shown.
Binary file not shown.
28
SumasenLibs/excel_lib/testdata/sample.py
vendored
28
SumasenLibs/excel_lib/testdata/sample.py
vendored
@ -1,28 +0,0 @@
|
|||||||
from sumaexcel import SumasenExcel
|
|
||||||
import logging
|
|
||||||
|
|
||||||
# 初期化
|
|
||||||
# 初期化
|
|
||||||
variables = {
|
|
||||||
"zekken_number":"5033",
|
|
||||||
"event_code":"FC岐阜",
|
|
||||||
"db":"rogdb",
|
|
||||||
"username":"admin",
|
|
||||||
"password":"admin123456",
|
|
||||||
"host":"localhost",
|
|
||||||
"port":"5432"
|
|
||||||
}
|
|
||||||
excel = SumasenExcel(document="test", variables=variables, docbase="./testdata")
|
|
||||||
|
|
||||||
logging.info("Excelファイル作成 step-1")
|
|
||||||
|
|
||||||
# シート初期化
|
|
||||||
ret = excel.make_report(variables=variables)
|
|
||||||
logging.info(f"Excelファイル作成 step-2 : ret={ret}")
|
|
||||||
if ret["status"]==True:
|
|
||||||
filepath=ret["filepath"]
|
|
||||||
logging.info(f"Excelファイル作成 : ret.filepath={filepath}")
|
|
||||||
else:
|
|
||||||
message = ret.get("message", "No message provided")
|
|
||||||
logging.error(f"Excelファイル作成失敗 : ret.message={message}")
|
|
||||||
|
|
||||||
26
SumasenLibs/excel_lib/testdata/test.ini
vendored
26
SumasenLibs/excel_lib/testdata/test.ini
vendored
@ -1,26 +0,0 @@
|
|||||||
[basic]
|
|
||||||
template_file=certificate_template.xlsx
|
|
||||||
doc_file=certificate_[zekken_number].xlsx
|
|
||||||
sections=section1
|
|
||||||
maxcol=10
|
|
||||||
column_width=3,5,16,16,16,16,16,8,8,12,3
|
|
||||||
|
|
||||||
[section1]
|
|
||||||
template_sheet=certificate
|
|
||||||
sheet_name=certificate
|
|
||||||
groups=group1,group2
|
|
||||||
fit_to_width=1
|
|
||||||
orientation=portrait
|
|
||||||
|
|
||||||
[section1.group1]
|
|
||||||
table_name=mv_entry_details
|
|
||||||
where=zekken_number='[zekken_number]' and event_name='[event_code]'
|
|
||||||
group_range=A1:J12
|
|
||||||
|
|
||||||
|
|
||||||
[section1.group2]
|
|
||||||
table_name=v_checkins_locations
|
|
||||||
where=zekken_number='[zekken_number]' and event_code='[event_code]'
|
|
||||||
sort=path_order
|
|
||||||
group_range=A13:J13
|
|
||||||
|
|
||||||
@ -1,69 +0,0 @@
|
|||||||
<?xml version="1.0"?>
|
|
||||||
<!DOCTYPE fontconfig SYSTEM "fonts.dtd">
|
|
||||||
<fontconfig>
|
|
||||||
<dir>/usr/share/fonts</dir>
|
|
||||||
|
|
||||||
<!-- デフォルトのサンセリフフォントをIPAexGothicに設定 -->
|
|
||||||
<match target="pattern">
|
|
||||||
<test qual="any" name="family">
|
|
||||||
<string>sans-serif</string>
|
|
||||||
</test>
|
|
||||||
<edit name="family" mode="assign" binding="same">
|
|
||||||
<string>IPAexGothic</string>
|
|
||||||
</edit>
|
|
||||||
</match>
|
|
||||||
|
|
||||||
<!-- デフォルトのセリフフォントをIPAexMinchoに設定 -->
|
|
||||||
<match target="pattern">
|
|
||||||
<test qual="any" name="family">
|
|
||||||
<string>serif</string>
|
|
||||||
</test>
|
|
||||||
<edit name="family" mode="assign" binding="same">
|
|
||||||
<string>IPAexMincho</string>
|
|
||||||
</edit>
|
|
||||||
</match>
|
|
||||||
|
|
||||||
<!-- MS Gothic の代替としてIPAexGothicを使用 -->
|
|
||||||
<match target="pattern">
|
|
||||||
<test name="family">
|
|
||||||
<string>MS Gothic</string>
|
|
||||||
</test>
|
|
||||||
<edit name="family" mode="assign" binding="same">
|
|
||||||
<string>IPAexGothic</string>
|
|
||||||
</edit>
|
|
||||||
</match>
|
|
||||||
|
|
||||||
<!-- MS Mincho の代替としてIPAexMinchoを使用 -->
|
|
||||||
<match target="pattern">
|
|
||||||
<test name="family">
|
|
||||||
<string>MS Mincho</string>
|
|
||||||
</test>
|
|
||||||
<edit name="family" mode="assign" binding="same">
|
|
||||||
<string>IPAexMincho</string>
|
|
||||||
</edit>
|
|
||||||
</match>
|
|
||||||
|
|
||||||
<!-- ビットマップフォントを無効化 -->
|
|
||||||
<match target="font">
|
|
||||||
<edit name="embeddedbitmap" mode="assign">
|
|
||||||
<bool>false</bool>
|
|
||||||
</edit>
|
|
||||||
</match>
|
|
||||||
|
|
||||||
<!-- フォントのヒンティング設定 -->
|
|
||||||
<match target="font">
|
|
||||||
<edit name="hintstyle" mode="assign">
|
|
||||||
<const>hintslight</const>
|
|
||||||
</edit>
|
|
||||||
<edit name="rgba" mode="assign">
|
|
||||||
<const>rgb</const>
|
|
||||||
</edit>
|
|
||||||
</match>
|
|
||||||
|
|
||||||
<!-- アンチエイリアス設定 -->
|
|
||||||
<match target="font">
|
|
||||||
<edit name="antialias" mode="assign">
|
|
||||||
<bool>true</bool>
|
|
||||||
</edit>
|
|
||||||
</match>
|
|
||||||
</fontconfig>
|
|
||||||
@ -53,14 +53,10 @@ INSTALLED_APPS = [
|
|||||||
'leaflet',
|
'leaflet',
|
||||||
'leaflet_admin_list',
|
'leaflet_admin_list',
|
||||||
'rog.apps.RogConfig',
|
'rog.apps.RogConfig',
|
||||||
'corsheaders', # added
|
|
||||||
'django_filters'
|
'django_filters'
|
||||||
]
|
]
|
||||||
|
|
||||||
MIDDLEWARE = [
|
MIDDLEWARE = [
|
||||||
'corsheaders.middleware.CorsMiddleware', # できるだけ上部に
|
|
||||||
'django.middleware.common.CommonMiddleware',
|
|
||||||
|
|
||||||
'django.middleware.security.SecurityMiddleware',
|
'django.middleware.security.SecurityMiddleware',
|
||||||
'django.contrib.sessions.middleware.SessionMiddleware',
|
'django.contrib.sessions.middleware.SessionMiddleware',
|
||||||
'django.middleware.common.CommonMiddleware',
|
'django.middleware.common.CommonMiddleware',
|
||||||
@ -72,47 +68,10 @@ MIDDLEWARE = [
|
|||||||
|
|
||||||
ROOT_URLCONF = 'config.urls'
|
ROOT_URLCONF = 'config.urls'
|
||||||
|
|
||||||
CORS_ALLOW_ALL_ORIGINS = True # 開発環境のみ
|
|
||||||
CORS_ALLOW_CREDENTIALS = True
|
|
||||||
|
|
||||||
CORS_ALLOWED_METHODS = [
|
|
||||||
'GET',
|
|
||||||
'POST',
|
|
||||||
'PUT',
|
|
||||||
'PATCH',
|
|
||||||
'DELETE',
|
|
||||||
'OPTIONS'
|
|
||||||
]
|
|
||||||
CORS_ALLOWED_HEADERS = [
|
|
||||||
'accept',
|
|
||||||
'accept-encoding',
|
|
||||||
'authorization',
|
|
||||||
'content-type',
|
|
||||||
'dnt',
|
|
||||||
'origin',
|
|
||||||
'user-agent',
|
|
||||||
'x-csrftoken',
|
|
||||||
'x-requested-with',
|
|
||||||
]
|
|
||||||
|
|
||||||
# 本番環境では以下のように制限する
|
|
||||||
CORS_ALLOWED_ORIGINS = [
|
|
||||||
"https://rogaining.sumasen.net",
|
|
||||||
"http://rogaining.sumasen.net",
|
|
||||||
]
|
|
||||||
|
|
||||||
# CSRFの設定
|
|
||||||
CSRF_TRUSTED_ORIGINS = [
|
|
||||||
"http://rogaining.sumasen.net",
|
|
||||||
"https://rogaining.sumasen.net",
|
|
||||||
]
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
TEMPLATES = [
|
TEMPLATES = [
|
||||||
{
|
{
|
||||||
'BACKEND': 'django.template.backends.django.DjangoTemplates',
|
'BACKEND': 'django.template.backends.django.DjangoTemplates',
|
||||||
'DIRS': [os.path.join(BASE_DIR, 'templates')],
|
'DIRS': [BASE_DIR / 'templates'],
|
||||||
'APP_DIRS': True,
|
'APP_DIRS': True,
|
||||||
'OPTIONS': {
|
'OPTIONS': {
|
||||||
'context_processors': [
|
'context_processors': [
|
||||||
@ -179,12 +138,10 @@ USE_TZ = True
|
|||||||
STATIC_URL = '/static/'
|
STATIC_URL = '/static/'
|
||||||
|
|
||||||
#STATIC_URL = '/static2/'
|
#STATIC_URL = '/static2/'
|
||||||
#STATIC_ROOT = BASE_DIR / "static"
|
STATIC_ROOT = BASE_DIR / "static"
|
||||||
STATIC_ROOT = os.path.join(BASE_DIR, 'static')
|
|
||||||
|
|
||||||
MEDIA_URL = '/media/'
|
MEDIA_URL = '/media/'
|
||||||
#MEDIA_ROOT = BASE_DIR / "media/"
|
MEDIA_ROOT = BASE_DIR / "media/"
|
||||||
MEDIA_ROOT = os.path.join(BASE_DIR, 'media')
|
|
||||||
|
|
||||||
#STATICFILES_DIRS = (os.path.join(BASE_DIR, "static2"),os.path.join(BASE_DIR, "media"))
|
#STATICFILES_DIRS = (os.path.join(BASE_DIR, "static2"),os.path.join(BASE_DIR, "media"))
|
||||||
|
|
||||||
@ -219,85 +176,14 @@ REST_FRAMEWORK = {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
#FRONTEND_URL = 'https://rogaining.intranet.sumasen.net' # フロントエンドのURLに適宜変更してください
|
# Email settings
|
||||||
FRONTEND_URL = 'https://rogaining.sumasen.net' # フロントエンドのURLに適宜変更してください
|
|
||||||
|
|
||||||
# この設定により、メールは実際には送信されず、代わりにコンソールに出力されます。
|
|
||||||
EMAIL_BACKEND = 'django.core.mail.backends.smtp.EmailBackend'
|
EMAIL_BACKEND = 'django.core.mail.backends.smtp.EmailBackend'
|
||||||
|
EMAIL_HOST = 'smtp-mail.outlook.com'
|
||||||
EMAIL_HOST = 'smtp.outlook.com'
|
|
||||||
EMAIL_PORT = 587
|
EMAIL_PORT = 587
|
||||||
EMAIL_USE_TLS = True
|
EMAIL_USE_TLS = True
|
||||||
EMAIL_HOST_USER = 'rogaining@gifuai.net'
|
EMAIL_HOST_USER = 'akira.miyata@gifuai.net'
|
||||||
EMAIL_HOST_PASSWORD = 'ctcpy9823"x~'
|
EMAIL_HOST_PASSWORD = 'SachikoMiyata@123'
|
||||||
DEFAULT_FROM_EMAIL = 'rogaining@gifuai.net'
|
|
||||||
|
|
||||||
APP_DOWNLOAD_LINK = 'https://apps.apple.com/jp/app/%E5%B2%90%E9%98%9C%E3%83%8A%E3%83%93/id6444221792'
|
# Optional: Set a default "from" address
|
||||||
ANDROID_DOWNLOAD_LINK = 'https://play.google.com/store/apps/details?id=com.dvox.gifunavi&hl=ja'
|
DEFAULT_FROM_EMAIL = 'info@gifuai.net'
|
||||||
|
|
||||||
SERVICE_NAME = '岐阜ナビ(岐阜ロゲのアプリ)'
|
|
||||||
|
|
||||||
# settings.py
|
|
||||||
DEFAULT_CHARSET = 'utf-8'
|
|
||||||
|
|
||||||
#REST_FRAMEWORK = {
|
|
||||||
# 'DEFAULT_RENDERER_CLASSES': [
|
|
||||||
# 'rest_framework.renderers.JSONRenderer',
|
|
||||||
# ],
|
|
||||||
# 'JSON_UNICODE_ESCAPE': False,
|
|
||||||
#}
|
|
||||||
|
|
||||||
LOGGING = {
|
|
||||||
'version': 1,
|
|
||||||
'disable_existing_loggers': False,
|
|
||||||
'formatters': {
|
|
||||||
'verbose': {
|
|
||||||
'format': '{levelname} {asctime} {module} {message}',
|
|
||||||
'style': '{',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
'handlers': {
|
|
||||||
#'file': {
|
|
||||||
# 'level': 'DEBUG',
|
|
||||||
# 'class': 'logging.FileHandler',
|
|
||||||
# 'filename': os.path.join(BASE_DIR, 'logs/debug.log'),
|
|
||||||
# 'formatter': 'verbose',
|
|
||||||
#},
|
|
||||||
'console': {
|
|
||||||
'level': 'DEBUG',
|
|
||||||
'class': 'logging.StreamHandler',
|
|
||||||
'formatter': 'verbose',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
'root': {
|
|
||||||
'handlers': ['console'],
|
|
||||||
'level': 'DEBUG',
|
|
||||||
},
|
|
||||||
'loggers': {
|
|
||||||
'django': {
|
|
||||||
'handlers': ['console'],
|
|
||||||
'level': 'INFO',
|
|
||||||
'propagate': False,
|
|
||||||
},
|
|
||||||
'django.request': {
|
|
||||||
'handlers': ['console'],
|
|
||||||
'level': 'DEBUG',
|
|
||||||
},
|
|
||||||
'rog': {
|
|
||||||
#'handlers': ['file','console'],
|
|
||||||
'handlers': ['console'],
|
|
||||||
'level': 'DEBUG',
|
|
||||||
'propagate': True,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
PASSWORD_HASHERS = [
|
|
||||||
'django.contrib.auth.hashers.PBKDF2PasswordHasher',
|
|
||||||
'django.contrib.auth.hashers.PBKDF2SHA1PasswordHasher',
|
|
||||||
'django.contrib.auth.hashers.Argon2PasswordHasher',
|
|
||||||
'django.contrib.auth.hashers.BCryptSHA256PasswordHasher',
|
|
||||||
]
|
|
||||||
|
|
||||||
BLACKLISTED_IPS = ['44.230.58.114'] # ブロックしたい IP アドレスをここにリストとして追加
|
|
||||||
|
|
||||||
|
|||||||
@ -18,21 +18,6 @@ from django.urls import path, include
|
|||||||
from django.conf import settings
|
from django.conf import settings
|
||||||
from django.conf.urls.static import static
|
from django.conf.urls.static import static
|
||||||
|
|
||||||
|
|
||||||
# debug_urlsビューをrogアプリケーションのviewsからインポート
|
|
||||||
from rog import views as rog_views
|
|
||||||
|
|
||||||
DEBUG = True
|
|
||||||
ALLOWED_HOSTS = ['rogaining.sumasen.net', 'localhost', '127.0.0.1']
|
|
||||||
|
|
||||||
# CORSの設定
|
|
||||||
CORS_ALLOW_ALL_ORIGINS = True
|
|
||||||
CORS_ALLOWED_ORIGINS = [
|
|
||||||
"http://rogaining.sumasen.net",
|
|
||||||
"http://localhost",
|
|
||||||
"http://127.0.0.1",
|
|
||||||
]
|
|
||||||
|
|
||||||
urlpatterns = [
|
urlpatterns = [
|
||||||
path('admin/', admin.site.urls),
|
path('admin/', admin.site.urls),
|
||||||
path('auth/', include('knox.urls')),
|
path('auth/', include('knox.urls')),
|
||||||
@ -42,8 +27,3 @@ urlpatterns = [
|
|||||||
admin.site.site_header = "ROGANING"
|
admin.site.site_header = "ROGANING"
|
||||||
admin.site.site_title = "Roganing Admin Portal"
|
admin.site.site_title = "Roganing Admin Portal"
|
||||||
admin.site.index_title = "Welcome to Roganing Portal"
|
admin.site.index_title = "Welcome to Roganing Portal"
|
||||||
|
|
||||||
# 開発環境での静的ファイル配信
|
|
||||||
if settings.DEBUG:
|
|
||||||
urlpatterns += static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)
|
|
||||||
urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
|
|
||||||
|
|||||||
@ -1,27 +0,0 @@
|
|||||||
[basic]
|
|
||||||
template_file=certificate_template.xlsx
|
|
||||||
doc_file=certificate_[zekken_number].xlsx
|
|
||||||
sections=section1
|
|
||||||
maxcol=10
|
|
||||||
column_width=3,5,16,16,16,20,16,8,8,12,3
|
|
||||||
output_path=media/reports/[event_code]
|
|
||||||
|
|
||||||
[section1]
|
|
||||||
template_sheet=certificate
|
|
||||||
sheet_name=certificate
|
|
||||||
groups=group1,group2
|
|
||||||
fit_to_width=1
|
|
||||||
orientation=portrait
|
|
||||||
|
|
||||||
[section1.group1]
|
|
||||||
table_name=mv_entry_details
|
|
||||||
where=zekken_number='[zekken_number]' and event_name='[event_code]'
|
|
||||||
group_range=A1:K15
|
|
||||||
|
|
||||||
|
|
||||||
[section1.group2]
|
|
||||||
table_name=v_checkins_locations
|
|
||||||
where=zekken_number='[zekken_number]' and event_code='[event_code]'
|
|
||||||
sort=path_order
|
|
||||||
group_range=A16:J16
|
|
||||||
|
|
||||||
Binary file not shown.
@ -1,77 +1,46 @@
|
|||||||
version: "3.9"
|
version: "3.9"
|
||||||
|
|
||||||
services:
|
services:
|
||||||
# postgres-db:
|
postgres-db:
|
||||||
# image: kartoza/postgis:12.0
|
image: kartoza/postgis:12.0
|
||||||
# ports:
|
|
||||||
# - 5432:5432
|
|
||||||
# volumes:
|
|
||||||
# - postgres_data:/var/lib/postgresql
|
|
||||||
# - ./custom-postgresql.conf:/etc/postgresql/12/main/postgresql.conf
|
|
||||||
# environment:
|
|
||||||
# - POSTGRES_USER=${POSTGRES_USER}
|
|
||||||
# - POSTGRES_PASS=${POSTGRES_PASS}
|
|
||||||
# - POSTGRES_DBNAME=${POSTGRES_DBNAME}
|
|
||||||
# - POSTGRES_MAX_CONNECTIONS=600
|
|
||||||
|
|
||||||
# restart: "on-failure"
|
|
||||||
# networks:
|
|
||||||
# - rog-api
|
|
||||||
|
|
||||||
api:
|
|
||||||
build:
|
|
||||||
context: .
|
|
||||||
dockerfile: Dockerfile.gdal
|
|
||||||
command: python3 manage.py runserver 0.0.0.0:8100
|
|
||||||
volumes:
|
|
||||||
- .:/app
|
|
||||||
ports:
|
|
||||||
- 8100:8100
|
|
||||||
env_file:
|
|
||||||
- .env
|
|
||||||
restart: "on-failure"
|
|
||||||
networks:
|
|
||||||
- rog-api
|
|
||||||
|
|
||||||
supervisor-web:
|
|
||||||
build:
|
|
||||||
context: .
|
|
||||||
dockerfile: Dockerfile.supervisor
|
|
||||||
volumes:
|
|
||||||
- type: bind
|
|
||||||
source: ./supervisor/html
|
|
||||||
target: /usr/share/nginx/html/supervisor
|
|
||||||
read_only: true
|
|
||||||
- type: bind
|
|
||||||
source: ./supervisor/nginx/default.conf
|
|
||||||
target: /etc/nginx/conf.d/default.conf
|
|
||||||
read_only: true
|
|
||||||
- type: volume
|
|
||||||
source: static_volume
|
|
||||||
target: /app/static
|
|
||||||
read_only: true
|
|
||||||
- type: volume
|
|
||||||
source: nginx_logs
|
|
||||||
target: /var/log/nginx
|
|
||||||
- type: bind
|
|
||||||
source: ./media
|
|
||||||
target: /usr/share/nginx/html/media
|
|
||||||
ports:
|
ports:
|
||||||
- "80:80"
|
- 5432:5432
|
||||||
depends_on:
|
volumes:
|
||||||
- api
|
- postgres_data:/var/lib/postgresql
|
||||||
|
- ./custom-postgresql.conf:/etc/postgresql/12/main/postgresql.conf
|
||||||
|
environment:
|
||||||
|
- POSTGRES_USER=${POSTGRES_USER}
|
||||||
|
- POSTGRES_PASS=${POSTGRES_PASS}
|
||||||
|
- POSTGRES_DBNAME=${POSTGRES_DBNAME}
|
||||||
|
- POSTGRES_MAX_CONNECTIONS=600
|
||||||
|
|
||||||
|
restart: "on-failure"
|
||||||
networks:
|
networks:
|
||||||
- rog-api
|
- rog-api
|
||||||
restart: always
|
|
||||||
|
api:
|
||||||
|
build:
|
||||||
|
context: .
|
||||||
|
dockerfile: Dockerfile.gdal
|
||||||
|
command: python3 manage.py runserver 0.0.0.0:8100
|
||||||
|
volumes:
|
||||||
|
- .:/app
|
||||||
|
ports:
|
||||||
|
- 8100:8100
|
||||||
|
env_file:
|
||||||
|
- .env
|
||||||
|
restart: "on-failure"
|
||||||
|
depends_on:
|
||||||
|
- postgres-db
|
||||||
|
networks:
|
||||||
|
- rog-api
|
||||||
|
#entrypoint: ["/app/wait-for.sh", "postgres-db:5432", "--", ""]
|
||||||
|
#command: python3 manage.py runserver 0.0.0.0:8100
|
||||||
|
|
||||||
networks:
|
networks:
|
||||||
rog-api:
|
rog-api:
|
||||||
driver: bridge
|
driver: bridge
|
||||||
|
|
||||||
volumes:
|
volumes:
|
||||||
postgres_data:
|
postgres_data:
|
||||||
geoserver-data:
|
geoserver-data:
|
||||||
static_volume:
|
|
||||||
nginx_logs:
|
|
||||||
@ -1,81 +0,0 @@
|
|||||||
version: "3.9"
|
|
||||||
|
|
||||||
services:
|
|
||||||
# postgres-db:
|
|
||||||
# image: kartoza/postgis:12.0
|
|
||||||
# ports:
|
|
||||||
# - 5432:5432
|
|
||||||
# volumes:
|
|
||||||
# - postgres_data:/var/lib/postgresql
|
|
||||||
# - ./custom-postgresql.conf:/etc/postgresql/12/main/postgresql.conf
|
|
||||||
# environment:
|
|
||||||
# - POSTGRES_USER=${POSTGRES_USER}
|
|
||||||
# - POSTGRES_PASS=${POSTGRES_PASS}
|
|
||||||
# - POSTGRES_DBNAME=${POSTGRES_DBNAME}
|
|
||||||
# - POSTGRES_MAX_CONNECTIONS=600
|
|
||||||
|
|
||||||
# restart: "on-failure"
|
|
||||||
# networks:
|
|
||||||
# - rog-api
|
|
||||||
|
|
||||||
api:
|
|
||||||
build:
|
|
||||||
context: .
|
|
||||||
dockerfile: Dockerfile.gdal
|
|
||||||
command: python3 manage.py runserver 0.0.0.0:8100
|
|
||||||
volumes:
|
|
||||||
- .:/app
|
|
||||||
ports:
|
|
||||||
- 8100:8100
|
|
||||||
env_file:
|
|
||||||
- .env
|
|
||||||
restart: "on-failure"
|
|
||||||
# depends_on:
|
|
||||||
# - postgres-db
|
|
||||||
networks:
|
|
||||||
- rog-api
|
|
||||||
#entrypoint: ["/app/wait-for.sh", "postgres-db:5432", "--", ""]
|
|
||||||
#command: python3 manage.py runserver 0.0.0.0:8100
|
|
||||||
|
|
||||||
supervisor-web:
|
|
||||||
build:
|
|
||||||
context: .
|
|
||||||
dockerfile: Dockerfile.supervisor
|
|
||||||
volumes:
|
|
||||||
- type: bind
|
|
||||||
source: /etc/letsencrypt
|
|
||||||
target: /etc/nginx/ssl
|
|
||||||
read_only: true
|
|
||||||
- type: bind
|
|
||||||
source: ./supervisor/html
|
|
||||||
target: /usr/share/nginx/html
|
|
||||||
read_only: true
|
|
||||||
- type: bind
|
|
||||||
source: ./supervisor/nginx/default.conf
|
|
||||||
target: /etc/nginx/conf.d/default.conf
|
|
||||||
read_only: true
|
|
||||||
- type: volume
|
|
||||||
source: static_volume
|
|
||||||
target: /app/static
|
|
||||||
read_only: true
|
|
||||||
- type: volume
|
|
||||||
source: nginx_logs
|
|
||||||
target: /var/log/nginx
|
|
||||||
ports:
|
|
||||||
- "80:80"
|
|
||||||
depends_on:
|
|
||||||
- api
|
|
||||||
networks:
|
|
||||||
- rog-api
|
|
||||||
restart: always
|
|
||||||
|
|
||||||
|
|
||||||
networks:
|
|
||||||
rog-api:
|
|
||||||
driver: bridge
|
|
||||||
|
|
||||||
volumes:
|
|
||||||
postgres_data:
|
|
||||||
geoserver-data:
|
|
||||||
static_volume:
|
|
||||||
nginx_logs:
|
|
||||||
@ -52,5 +52,4 @@ releasenotes/build
|
|||||||
hosts
|
hosts
|
||||||
*.retry
|
*.retry
|
||||||
#Vagrantfiles, since we are using docker
|
#Vagrantfiles, since we are using docker
|
||||||
Vagrantfile.*
|
Vagrantfile.*
|
||||||
media/
|
|
||||||
@ -1 +0,0 @@
|
|||||||
pg_dump: error: connection to database "rogdb" failed: FATAL: Peer authentication failed for user "postgres"
|
|
||||||
@ -1,10 +0,0 @@
|
|||||||
#!/bin/sh
|
|
||||||
|
|
||||||
# Collect static files
|
|
||||||
python manage.py collectstatic --noinput
|
|
||||||
|
|
||||||
# Apply database migrations
|
|
||||||
python manage.py migrate
|
|
||||||
|
|
||||||
# Start Gunicorn
|
|
||||||
exec "$@"
|
|
||||||
32
nginx.conf
32
nginx.conf
@ -26,33 +26,29 @@ http {
|
|||||||
#gzip on;
|
#gzip on;
|
||||||
|
|
||||||
server {
|
server {
|
||||||
listen 80;
|
listen 80;
|
||||||
server_name localhost;
|
server_name localhost;
|
||||||
|
|
||||||
# 静的ファイルの提供
|
|
||||||
location /static/ {
|
location /static/ {
|
||||||
alias /app/static/;
|
alias /app/static/;
|
||||||
}
|
}
|
||||||
|
|
||||||
# スーパーバイザー Web アプリケーション
|
location /media/ {
|
||||||
|
alias /app/media/;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
location / {
|
location / {
|
||||||
root /usr/share/nginx/html;
|
proxy_pass http://app:8000;
|
||||||
index index.html;
|
|
||||||
try_files $uri $uri/ /index.html;
|
|
||||||
}
|
|
||||||
|
|
||||||
# Django API プロキシ
|
|
||||||
location /api/ {
|
|
||||||
proxy_pass http://api:8000;
|
|
||||||
proxy_set_header Host $host;
|
proxy_set_header Host $host;
|
||||||
proxy_set_header X-Real-IP $remote_addr;
|
proxy_set_header X-Real-IP $remote_addr;
|
||||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||||
proxy_set_header X-Forwarded-Proto $scheme;
|
proxy_set_header X-Forwarded-Proto $scheme;
|
||||||
}
|
}
|
||||||
|
|
||||||
error_page 500 502 503 504 /50x.html;
|
error_page 500 502 503 504 /50x.html;
|
||||||
location = /50x.html {
|
location = /50x.html {
|
||||||
root /usr/share/nginx/html;
|
root /usr/share/nginx/html;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@ -31,7 +31,7 @@ matplotlib==3.5.0
|
|||||||
mccabe==0.6.1
|
mccabe==0.6.1
|
||||||
munch==2.5.0
|
munch==2.5.0
|
||||||
mypy-extensions==0.4.3
|
mypy-extensions==0.4.3
|
||||||
numpy==1.21.4
|
numpy==1.26.2
|
||||||
packaging==21.3
|
packaging==21.3
|
||||||
pandas==1.3.4
|
pandas==1.3.4
|
||||||
pathspec==0.9.0
|
pathspec==0.9.0
|
||||||
@ -46,11 +46,13 @@ pyparsing==3.0.6
|
|||||||
pyproj==3.3.0
|
pyproj==3.3.0
|
||||||
python-dateutil==2.8.2
|
python-dateutil==2.8.2
|
||||||
pytz==2021.3
|
pytz==2021.3
|
||||||
rasterio==1.2.10
|
#rasterio==1.2.10 Akira
|
||||||
|
rasterio==1.3.10
|
||||||
regex==2021.11.10
|
regex==2021.11.10
|
||||||
requests==2.26.0
|
requests==2.26.0
|
||||||
Rtree==0.9.7
|
Rtree==0.9.7
|
||||||
scipy==1.7.3
|
#scipy==1.7.3
|
||||||
|
scipy==1.10.1
|
||||||
seaborn==0.11.2
|
seaborn==0.11.2
|
||||||
setuptools-scm==6.3.2
|
setuptools-scm==6.3.2
|
||||||
Shapely==1.8.0
|
Shapely==1.8.0
|
||||||
@ -65,4 +67,3 @@ django-extra-fields==3.0.2
|
|||||||
django-phonenumber-field==6.1.0
|
django-phonenumber-field==6.1.0
|
||||||
django-rest-knox==4.2.0
|
django-rest-knox==4.2.0
|
||||||
dj-database-url==2.0.0
|
dj-database-url==2.0.0
|
||||||
django-cors-headers==4.3.0
|
|
||||||
|
|||||||
BIN
rog/.DS_Store
vendored
BIN
rog/.DS_Store
vendored
Binary file not shown.
1228
rog/admin.py
1228
rog/admin.py
File diff suppressed because it is too large
Load Diff
@ -2,19 +2,12 @@ from django.conf import settings
|
|||||||
#from django.contrib.auth import get_user_model
|
#from django.contrib.auth import get_user_model
|
||||||
from .models import CustomUser
|
from .models import CustomUser
|
||||||
from django.contrib.auth.backends import ModelBackend
|
from django.contrib.auth.backends import ModelBackend
|
||||||
from django.contrib.auth import get_user_model
|
|
||||||
from django.contrib.auth.hashers import check_password
|
|
||||||
import logging
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class EmailOrUsernameModelBackend(ModelBackend):
|
class EmailOrUsernameModelBackend(ModelBackend):
|
||||||
"""
|
"""
|
||||||
This is a ModelBacked that allows authentication
|
This is a ModelBacked that allows authentication
|
||||||
with either a username or an email address.
|
with either a username or an email address.
|
||||||
|
|
||||||
"""
|
|
||||||
"""
|
"""
|
||||||
def authenticate(self, username=None, password=None):
|
def authenticate(self, username=None, password=None):
|
||||||
if '@' in username:
|
if '@' in username:
|
||||||
@ -32,35 +25,4 @@ class EmailOrUsernameModelBackend(ModelBackend):
|
|||||||
try:
|
try:
|
||||||
return CustomUser.objects.get(pk=username)
|
return CustomUser.objects.get(pk=username)
|
||||||
except get_user_model().DoesNotExist:
|
except get_user_model().DoesNotExist:
|
||||||
return None
|
return None
|
||||||
"""
|
|
||||||
|
|
||||||
def authenticate(self, request, username=None, password=None, **kwargs):
|
|
||||||
if '@' in username:
|
|
||||||
kwargs = {'email': username}
|
|
||||||
else:
|
|
||||||
kwargs = {'username': username}
|
|
||||||
try:
|
|
||||||
user = CustomUser.objects.get(**kwargs)
|
|
||||||
if check_password(password, user.password):
|
|
||||||
logger.info(f"User authenticated successfully: {username}")
|
|
||||||
return user
|
|
||||||
else:
|
|
||||||
logger.warning(f"Password mismatch for user: {username}")
|
|
||||||
except CustomUser.DoesNotExist:
|
|
||||||
logger.warning(f"User does not exist: {username}")
|
|
||||||
except Exception as e:
|
|
||||||
logger.error(f"Authentication error for {username}: {str(e)}")
|
|
||||||
return None
|
|
||||||
|
|
||||||
def get_user(self, user_id):
|
|
||||||
try:
|
|
||||||
user = CustomUser.objects.get(pk=user_id)
|
|
||||||
logger.info(f"User retrieved: {user.username or user.email}")
|
|
||||||
return user
|
|
||||||
except CustomUser.DoesNotExist:
|
|
||||||
logger.warning(f"User with id {user_id} does not exist")
|
|
||||||
return None
|
|
||||||
except Exception as e:
|
|
||||||
logger.error(f"Error retrieving user with id {user_id}: {str(e)}")
|
|
||||||
return None
|
|
||||||
@ -1,7 +0,0 @@
|
|||||||
from django import forms
|
|
||||||
from .models import NewEvent2
|
|
||||||
|
|
||||||
class CSVUploadForm(forms.Form):
|
|
||||||
event = forms.ModelChoiceField(queryset=NewEvent2.objects.all(), label="イベント選択")
|
|
||||||
csv_file = forms.FileField(label="CSVファイル")
|
|
||||||
|
|
||||||
@ -1,40 +0,0 @@
|
|||||||
4019,関ケ原2,Best Wishes,ソロ女子-5時間,pbkdf2_sha256$260000$RPvncicp11ENXxwpcpMXi1$9e/fKcfwaX3sJ91q9S70KWQcrNlraliguiHjF/UCW/I=
|
|
||||||
4010,関ケ原2,まつげん,ソロ女子-5時間,pbkdf2_sha256$260000$LMvH0KtHeHbCuuUZ5n88VZ$Lnsqs/u45QKoFN6lUdqC79nIMz5LwaKWMpmX/0aEXa8=
|
|
||||||
4021,大垣3,まつげん,ソロ女子-5時間,pbkdf2_sha256$260000$LMvH0KtHeHbCuuUZ5n88VZ$Lnsqs/u45QKoFN6lUdqC79nIMz5LwaKWMpmX/0aEXa8=
|
|
||||||
5,関ケ原2,てすとあきら1,ソロ男子-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
3003,関ケ原2,てすとあきら1,ソロ男子-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
3115,関ケ原2,Best Wishes,ソロ男子-5時間,pbkdf2_sha256$260000$tlNrgHyqDtfbM9f3GLv5G1$jRcR/ieTB174TZ9jW7obCBUMpyz86aywqDKw3VmhVQQ=
|
|
||||||
1010,大垣3,ハヤノテスト,一般-5時間,pbkdf2_sha256$260000$IeGmRkkUkwXXc1zO9oxvCe$ijnJTH7xhwidit+uCggSgjj/7g/vMK539IpOMA5GlnM=
|
|
||||||
1012,大垣3,てすとあきら1,一般-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
1014,各務原2,てすとあきら1,一般-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
1018,下呂2,てすとあきら1,一般-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
1024,関ケ原2,てすとあきら1,一般-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
1026,美濃加茂2,てすとあきら1,一般-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
1028,多治見2,てすとあきら1,一般-5時間,pbkdf2_sha256$260000$0GY5pt5V127jGd8HkkEort$8ZL0eY2qTZHydyzUUN5LNKZnmmibfu1x3QQ/7rJX1Vc=
|
|
||||||
3006,関ケ原2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3009,養老2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3011,郡上2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3013,大垣3,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3015,各務原2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3017,多治見2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3019,下呂2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3021,高山2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
3023,美濃加茂2,山本哲也,ソロ男子-5時間,pbkdf2_sha256$260000$EkYrRHZwKunjO4jiHvxyB2$kYGN0STzV9c70IKAIxK1Ija3K1y90+ote0HDTP+iSPw=
|
|
||||||
4008,下呂2,GO!GO!YOKO,ソロ女子-5時間,pbkdf2_sha256$260000$tuv8ajw2VSmCooIxNHJhdD$m7q0fqPIsAs7L9uubt+PUVsmexwpJPXPCgVs9GjY12c=
|
|
||||||
3121,関ケ原2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3126,大垣3,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$7KsSngw2Ho719jpXsOrC8v$jfHFxglG/L0htA13t01LAy91dS+FnlAZubg6Lmd/m2Y=
|
|
||||||
3128,多治見2,MASA,ソロ男子-5時間,pbkdf2_sha256$260000$qpaSbqryD4f5bZaY893Ug4$Gk8XuqsJbSkX9Hxrl/xg9LtjM8JQkpgNkpbbNzTmhzY=
|
|
||||||
3124,関ケ原2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$7KsSngw2Ho719jpXsOrC8v$jfHFxglG/L0htA13t01LAy91dS+FnlAZubg6Lmd/m2Y=
|
|
||||||
3132,各務原2,岐阜市イイとこあるある探検隊,ソロ男子-5時間,pbkdf2_sha256$260000$QWc5BpSBUbkUwP9UlIzyE5$do+VKkH8mNibg6PJDsm6AJ/VMFh3NWdzwZ9IQW/26xA=
|
|
||||||
3135,大垣3,akira,ソロ男子-5時間,pbkdf2_sha256$260000$mmM2N8sSE84YaNNuDzQKxb$ox9U6rdgZq4ANzi4NizskphZWIrf7o2+JEfvC4wcn7U=
|
|
||||||
3137,関ケ原2,akira,ソロ男子-5時間,pbkdf2_sha256$260000$mmM2N8sSE84YaNNuDzQKxb$ox9U6rdgZq4ANzi4NizskphZWIrf7o2+JEfvC4wcn7U=
|
|
||||||
3139,養老2,akira,ソロ男子-5時間,pbkdf2_sha256$260000$mmM2N8sSE84YaNNuDzQKxb$ox9U6rdgZq4ANzi4NizskphZWIrf7o2+JEfvC4wcn7U=
|
|
||||||
3073,養老2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3075,高山2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3077,郡上2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3081,美濃加茂2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3083,多治見2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3085,各務原2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3079,下呂2,yamadeus,ソロ男子-5時間,pbkdf2_sha256$260000$sCLRTCAxQIClyDmvfbMDm0$cU3dSGTPwKHl8T3EBZ6R19oZJGkadD48pKqywAhtJOk=
|
|
||||||
3093,関ケ原2,岐阜愛,ソロ男子-5時間,pbkdf2_sha256$260000$LFOINdd30aKaXoT9CNYY8A$eoAzV10+gp+tufabtcFOx6uoOktZUngzzDJ0WWs/v24=
|
|
||||||
3099,高山2,岐阜愛,ソロ男子-5時間,pbkdf2_sha256$260000$LFOINdd30aKaXoT9CNYY8A$eoAzV10+gp+tufabtcFOx6uoOktZUngzzDJ0WWs/v24=
|
|
||||||
|
@ -1,3 +0,0 @@
|
|||||||
from .ip_blocking import IPBlockingMiddleware
|
|
||||||
|
|
||||||
__all__ = ['IPBlockingMiddleware']
|
|
||||||
@ -1,42 +0,0 @@
|
|||||||
from django.core.exceptions import PermissionDenied
|
|
||||||
from django.core.cache import cache
|
|
||||||
from django.conf import settings
|
|
||||||
|
|
||||||
class IPBlockingMiddleware:
|
|
||||||
def __init__(self, get_response):
|
|
||||||
self.get_response = get_response
|
|
||||||
# 事前にブロックする IP アドレスのリスト
|
|
||||||
self.blacklisted_ips = getattr(settings, 'BLACKLISTED_IPS', [])
|
|
||||||
|
|
||||||
def __call__(self, request):
|
|
||||||
ip = self.get_client_ip(request)
|
|
||||||
|
|
||||||
# キャッシュからブロックリストを取得
|
|
||||||
blocked_ips = cache.get('blocked_ips', set())
|
|
||||||
|
|
||||||
# 事前にブロックされた IP またはキャッシュ内のブロックされた IP をチェック
|
|
||||||
if ip in self.blacklisted_ips or ip in blocked_ips:
|
|
||||||
raise PermissionDenied
|
|
||||||
|
|
||||||
# 不正アクセスの検出ロジックをここに実装
|
|
||||||
if self.is_suspicious(ip):
|
|
||||||
blocked_ips.add(ip)
|
|
||||||
cache.set('blocked_ips', blocked_ips, timeout=3600) # 1時間ブロック
|
|
||||||
raise PermissionDenied
|
|
||||||
|
|
||||||
response = self.get_response(request)
|
|
||||||
return response
|
|
||||||
|
|
||||||
def is_suspicious(self, ip):
|
|
||||||
request_count = cache.get(f'request_count_{ip}', 0)
|
|
||||||
cache.set(f'request_count_{ip}', request_count + 1, timeout=60)
|
|
||||||
return request_count > 100 # 1分間に100回以上のリクエストがあれば不審と判断
|
|
||||||
|
|
||||||
def get_client_ip(self, request):
|
|
||||||
x_forwarded_for = request.META.get('HTTP_X_FORWARDED_FOR')
|
|
||||||
if x_forwarded_for:
|
|
||||||
ip = x_forwarded_for.split(',')[0]
|
|
||||||
else:
|
|
||||||
ip = request.META.get('REMOTE_ADDR')
|
|
||||||
return ip
|
|
||||||
|
|
||||||
@ -1,148 +0,0 @@
|
|||||||
"""
|
|
||||||
このコードは永栄コードをNoufferコードに統合するための一時変換コードです。
|
|
||||||
一旦、完全にマイグレーションでき、ランキングや走行履歴が完成したら、不要になります。
|
|
||||||
"""
|
|
||||||
import psycopg2
|
|
||||||
from PIL import Image
|
|
||||||
import PIL.ExifTags
|
|
||||||
from datetime import datetime
|
|
||||||
import os
|
|
||||||
|
|
||||||
def get_gps_from_image(image_path):
|
|
||||||
"""
|
|
||||||
画像ファイルからGPS情報を抽出する
|
|
||||||
Returns: (latitude, longitude) または取得できない場合は (None, None)
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
with Image.open(image_path) as img:
|
|
||||||
exif = {
|
|
||||||
PIL.ExifTags.TAGS[k]: v
|
|
||||||
for k, v in img._getexif().items()
|
|
||||||
if k in PIL.ExifTags.TAGS
|
|
||||||
}
|
|
||||||
|
|
||||||
if 'GPSInfo' in exif:
|
|
||||||
gps_info = exif['GPSInfo']
|
|
||||||
|
|
||||||
# 緯度の計算
|
|
||||||
lat = gps_info[2]
|
|
||||||
lat = lat[0] + lat[1]/60 + lat[2]/3600
|
|
||||||
if gps_info[1] == 'S':
|
|
||||||
lat = -lat
|
|
||||||
|
|
||||||
# 経度の計算
|
|
||||||
lon = gps_info[4]
|
|
||||||
lon = lon[0] + lon[1]/60 + lon[2]/3600
|
|
||||||
if gps_info[3] == 'W':
|
|
||||||
lon = -lon
|
|
||||||
|
|
||||||
return lat, lon
|
|
||||||
except Exception as e:
|
|
||||||
print(f"GPS情報の抽出に失敗: {e}")
|
|
||||||
|
|
||||||
return None, None
|
|
||||||
|
|
||||||
def migrate_data():
|
|
||||||
# コンテナ環境用の接続情報
|
|
||||||
source_db = {
|
|
||||||
'dbname': 'gifuroge',
|
|
||||||
'user': 'admin', # 環境に合わせて変更
|
|
||||||
'password': 'admin123456', # 環境に合わせて変更
|
|
||||||
'host': 'localhost', # Dockerのサービス名
|
|
||||||
'port': '5432'
|
|
||||||
}
|
|
||||||
|
|
||||||
target_db = {
|
|
||||||
'dbname': 'rogdb',
|
|
||||||
'user': 'admin', # 環境に合わせて変更
|
|
||||||
'password': 'admin123456', # 環境に合わせて変更
|
|
||||||
'host': 'localhost', # Dockerのサービス名
|
|
||||||
'port': '5432'
|
|
||||||
}
|
|
||||||
|
|
||||||
source_conn = None
|
|
||||||
target_conn = None
|
|
||||||
source_cur = None
|
|
||||||
target_cur = None
|
|
||||||
|
|
||||||
try:
|
|
||||||
print("ソースDBへの接続を試みています...")
|
|
||||||
source_conn = psycopg2.connect(**source_db)
|
|
||||||
source_cur = source_conn.cursor()
|
|
||||||
print("ソースDBへの接続が成功しました")
|
|
||||||
|
|
||||||
print("ターゲットDBへの接続を試みています...")
|
|
||||||
target_conn = psycopg2.connect(**target_db)
|
|
||||||
target_cur = target_conn.cursor()
|
|
||||||
print("ターゲットDBへの接続が成功しました")
|
|
||||||
|
|
||||||
print("データの取得を開始します...")
|
|
||||||
source_cur.execute("""
|
|
||||||
SELECT serial_number, zekken_number, event_code, cp_number, image_address,
|
|
||||||
goal_time, late_point, create_at, create_user,
|
|
||||||
update_at, update_user, buy_flag, colabo_company_memo
|
|
||||||
FROM gps_information
|
|
||||||
""")
|
|
||||||
|
|
||||||
rows = source_cur.fetchall()
|
|
||||||
print(f"取得したレコード数: {len(rows)}")
|
|
||||||
|
|
||||||
processed_count = 0
|
|
||||||
for row in rows:
|
|
||||||
(serial_number, zekken_number, event_code, cp_number, image_address,
|
|
||||||
goal_time, late_point, create_at, create_user,
|
|
||||||
update_at, update_user, buy_flag, colabo_company_memo) = row
|
|
||||||
|
|
||||||
latitude, longitude = None, None
|
|
||||||
if image_address and os.path.exists(image_address):
|
|
||||||
latitude, longitude = get_gps_from_image(image_address)
|
|
||||||
|
|
||||||
target_cur.execute("""
|
|
||||||
INSERT INTO gps_checkins (
|
|
||||||
path_order, zekken_number, event_code, cp_number,
|
|
||||||
lattitude, longitude, image_address,
|
|
||||||
image_receipt, image_QR, validate_location,
|
|
||||||
goal_time, late_point, create_at,
|
|
||||||
create_user, update_at, update_user,
|
|
||||||
buy_flag, colabo_company_memo, points
|
|
||||||
) VALUES (
|
|
||||||
%s, %s, %s, %s, %s, %s, %s, %s, %s, %s,
|
|
||||||
%s, %s, %s, %s, %s, %s, %s, %s, %s
|
|
||||||
)
|
|
||||||
""", (
|
|
||||||
serial_number,
|
|
||||||
zekken_number, event_code, cp_number,
|
|
||||||
latitude, longitude, image_address,
|
|
||||||
True, True, True,
|
|
||||||
goal_time, late_point, create_at,
|
|
||||||
create_user, update_at, update_user,
|
|
||||||
buy_flag if buy_flag is not None else False,
|
|
||||||
colabo_company_memo if colabo_company_memo else '',
|
|
||||||
0
|
|
||||||
))
|
|
||||||
|
|
||||||
processed_count += 1
|
|
||||||
if processed_count % 100 == 0:
|
|
||||||
print(f"処理済みレコード数: {processed_count}")
|
|
||||||
|
|
||||||
target_conn.commit()
|
|
||||||
print(f"移行完了: {processed_count}件のレコードを処理しました")
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
print(f"エラーが発生しました: {e}")
|
|
||||||
if target_conn:
|
|
||||||
target_conn.rollback()
|
|
||||||
|
|
||||||
finally:
|
|
||||||
if source_cur:
|
|
||||||
source_cur.close()
|
|
||||||
if target_cur:
|
|
||||||
target_cur.close()
|
|
||||||
if source_conn:
|
|
||||||
source_conn.close()
|
|
||||||
if target_conn:
|
|
||||||
target_conn.close()
|
|
||||||
print("すべての接続をクローズしました")
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
migrate_data()
|
|
||||||
18
rog/migrations/0033_alter_templocation_sub_loc_id.py
Normal file
18
rog/migrations/0033_alter_templocation_sub_loc_id.py
Normal file
@ -0,0 +1,18 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-09-09 13:18
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0032_alter_location_sub_loc_id'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='templocation',
|
||||||
|
name='sub_loc_id',
|
||||||
|
field=models.CharField(blank=True, max_length=2048, null=True, verbose_name='Sub location id'),
|
||||||
|
),
|
||||||
|
]
|
||||||
18
rog/migrations/0034_alter_customuser_email.py
Normal file
18
rog/migrations/0034_alter_customuser_email.py
Normal file
@ -0,0 +1,18 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-10-06 10:51
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0033_alter_templocation_sub_loc_id'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='customuser',
|
||||||
|
name='email',
|
||||||
|
field=models.CharField(max_length=255, verbose_name='user name'),
|
||||||
|
),
|
||||||
|
]
|
||||||
18
rog/migrations/0035_alter_customuser_email.py
Normal file
18
rog/migrations/0035_alter_customuser_email.py
Normal file
@ -0,0 +1,18 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-10-06 10:52
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0034_alter_customuser_email'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='customuser',
|
||||||
|
name='email',
|
||||||
|
field=models.EmailField(max_length=254, unique=True, verbose_name='user name'),
|
||||||
|
),
|
||||||
|
]
|
||||||
18
rog/migrations/0036_alter_customuser_email.py
Normal file
18
rog/migrations/0036_alter_customuser_email.py
Normal file
@ -0,0 +1,18 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-10-06 11:01
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0035_alter_customuser_email'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='customuser',
|
||||||
|
name='email',
|
||||||
|
field=models.CharField(max_length=255, unique=True, verbose_name='Email'),
|
||||||
|
),
|
||||||
|
]
|
||||||
18
rog/migrations/0037_customuser_is_rogaining.py
Normal file
18
rog/migrations/0037_customuser_is_rogaining.py
Normal file
@ -0,0 +1,18 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-10-06 13:42
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0036_alter_customuser_email'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AddField(
|
||||||
|
model_name='customuser',
|
||||||
|
name='is_rogaining',
|
||||||
|
field=models.BooleanField(default=False),
|
||||||
|
),
|
||||||
|
]
|
||||||
28
rog/migrations/0038_auto_20221016_1950.py
Normal file
28
rog/migrations/0038_auto_20221016_1950.py
Normal file
@ -0,0 +1,28 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-10-16 10:50
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0037_customuser_is_rogaining'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AddField(
|
||||||
|
model_name='customuser',
|
||||||
|
name='event_code',
|
||||||
|
field=models.CharField(blank=True, max_length=255, null=True, verbose_name='Event Code'),
|
||||||
|
),
|
||||||
|
migrations.AddField(
|
||||||
|
model_name='customuser',
|
||||||
|
name='team_name',
|
||||||
|
field=models.CharField(blank=True, max_length=255, null=True, verbose_name='Team Name'),
|
||||||
|
),
|
||||||
|
migrations.AddField(
|
||||||
|
model_name='customuser',
|
||||||
|
name='zekken_number',
|
||||||
|
field=models.CharField(blank=True, max_length=255, null=True, verbose_name='Zekken Number'),
|
||||||
|
),
|
||||||
|
]
|
||||||
27
rog/migrations/0039_goalimages.py
Normal file
27
rog/migrations/0039_goalimages.py
Normal file
@ -0,0 +1,27 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-10-17 11:39
|
||||||
|
|
||||||
|
from django.conf import settings
|
||||||
|
from django.db import migrations, models
|
||||||
|
import django.db.models.deletion
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0038_auto_20221016_1950'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.CreateModel(
|
||||||
|
name='GoalImages',
|
||||||
|
fields=[
|
||||||
|
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
||||||
|
('goalimage', models.FileField(blank=True, null=True, upload_to='%y%m%d')),
|
||||||
|
('goaltime', models.DateTimeField(verbose_name='Goal time')),
|
||||||
|
('team_name', models.CharField(max_length=255, verbose_name='Team name')),
|
||||||
|
('event_code', models.CharField(max_length=255, verbose_name='event code')),
|
||||||
|
('cp_number', models.IntegerField(verbose_name='CP numner')),
|
||||||
|
('user', models.ForeignKey(on_delete=django.db.models.deletion.DO_NOTHING, to=settings.AUTH_USER_MODEL)),
|
||||||
|
],
|
||||||
|
),
|
||||||
|
]
|
||||||
32
rog/migrations/0040_auto_20221105_1939.py
Normal file
32
rog/migrations/0040_auto_20221105_1939.py
Normal file
@ -0,0 +1,32 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2022-11-05 10:39
|
||||||
|
|
||||||
|
from django.conf import settings
|
||||||
|
from django.db import migrations, models
|
||||||
|
import django.db.models.deletion
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0039_goalimages'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='goalimages',
|
||||||
|
name='goalimage',
|
||||||
|
field=models.FileField(blank=True, null=True, upload_to='goals/%y%m%d'),
|
||||||
|
),
|
||||||
|
migrations.CreateModel(
|
||||||
|
name='CheckinImages',
|
||||||
|
fields=[
|
||||||
|
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
||||||
|
('checkinimage', models.FileField(blank=True, null=True, upload_to='checkin/%y%m%d')),
|
||||||
|
('checkintime', models.DateTimeField(verbose_name='Goal time')),
|
||||||
|
('team_name', models.CharField(max_length=255, verbose_name='Team name')),
|
||||||
|
('event_code', models.CharField(max_length=255, verbose_name='event code')),
|
||||||
|
('cp_number', models.IntegerField(verbose_name='CP numner')),
|
||||||
|
('user', models.ForeignKey(on_delete=django.db.models.deletion.DO_NOTHING, to=settings.AUTH_USER_MODEL)),
|
||||||
|
],
|
||||||
|
),
|
||||||
|
]
|
||||||
49
rog/migrations/0041_auto_20230526_1724.py
Normal file
49
rog/migrations/0041_auto_20230526_1724.py
Normal file
@ -0,0 +1,49 @@
|
|||||||
|
# Generated by Django 3.2.9 on 2023-05-26 08:24
|
||||||
|
|
||||||
|
import django.contrib.postgres.indexes
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('rog', '0040_auto_20221105_1939'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.DeleteModel(
|
||||||
|
name='JpnAdminPerf',
|
||||||
|
),
|
||||||
|
migrations.DeleteModel(
|
||||||
|
name='JpnSubPerf',
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='location',
|
||||||
|
name='category',
|
||||||
|
field=models.CharField(blank=True, db_index=True, max_length=2048, null=True, verbose_name='Category'),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='location',
|
||||||
|
name='event_active',
|
||||||
|
field=models.BooleanField(db_index=True, default=True, verbose_name='Is Event active'),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='location',
|
||||||
|
name='event_name',
|
||||||
|
field=models.CharField(blank=True, db_index=True, max_length=2048, null=True, verbose_name='Event name'),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='location',
|
||||||
|
name='group',
|
||||||
|
field=models.CharField(blank=True, db_index=True, max_length=2048, null=True, verbose_name='Group'),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='location',
|
||||||
|
name='location_id',
|
||||||
|
field=models.IntegerField(blank=True, db_index=True, null=True, verbose_name='Location id'),
|
||||||
|
),
|
||||||
|
migrations.AddIndex(
|
||||||
|
model_name='location',
|
||||||
|
index=django.contrib.postgres.indexes.GistIndex(fields=['geom'], name='rog_locatio_geom_4793cc_gist'),
|
||||||
|
),
|
||||||
|
]
|
||||||
@ -1,274 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-07-24 05:31
|
|
||||||
|
|
||||||
import datetime
|
|
||||||
from django.conf import settings
|
|
||||||
import django.contrib.postgres.indexes
|
|
||||||
from django.db import migrations, models
|
|
||||||
import django.db.models.deletion
|
|
||||||
import django.utils.timezone
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0032_alter_location_sub_loc_id'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='Category',
|
|
||||||
fields=[
|
|
||||||
('category_name', models.CharField(max_length=255, primary_key=True, serialize=False)),
|
|
||||||
('category_number', models.IntegerField(unique=True)),
|
|
||||||
('duration', models.DurationField(default=datetime.timedelta(seconds=18000))),
|
|
||||||
('num_of_member', models.IntegerField(default=1)),
|
|
||||||
('family', models.BooleanField(default=False)),
|
|
||||||
('female', models.BooleanField(default=False)),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='CheckinImages',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('checkinimage', models.FileField(blank=True, null=True, upload_to='checkin/%y%m%d')),
|
|
||||||
('checkintime', models.DateTimeField(verbose_name='Goal time')),
|
|
||||||
('team_name', models.CharField(max_length=255, verbose_name='Team name')),
|
|
||||||
('event_code', models.CharField(max_length=255, verbose_name='event code')),
|
|
||||||
('cp_number', models.IntegerField(verbose_name='CP numner')),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='Entry',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('date', models.DateTimeField()),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='EntryMember',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('is_temporary', models.BooleanField(default=False)),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='GoalImages',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('goalimage', models.FileField(blank=True, null=True, upload_to='goals/%y%m%d')),
|
|
||||||
('goaltime', models.DateTimeField(verbose_name='Goal time')),
|
|
||||||
('team_name', models.CharField(max_length=255, verbose_name='Team name')),
|
|
||||||
('event_code', models.CharField(max_length=255, verbose_name='event code')),
|
|
||||||
('cp_number', models.IntegerField(verbose_name='CP numner')),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='Member',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('is_temporary', models.BooleanField(default=False)),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='NewEvent',
|
|
||||||
fields=[
|
|
||||||
('event_name', models.CharField(max_length=255, primary_key=True, serialize=False)),
|
|
||||||
('start_datetime', models.DateTimeField(default=django.utils.timezone.now)),
|
|
||||||
('end_datetime', models.DateTimeField()),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='Team',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('zekken_number', models.CharField(max_length=255, unique=True)),
|
|
||||||
('team_name', models.CharField(max_length=255)),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='TempUser',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('email', models.EmailField(max_length=254, unique=True)),
|
|
||||||
('password', models.CharField(max_length=128)),
|
|
||||||
('is_rogaining', models.BooleanField(default=False)),
|
|
||||||
('zekken_number', models.CharField(blank=True, max_length=255, null=True)),
|
|
||||||
('event_code', models.CharField(blank=True, max_length=255, null=True)),
|
|
||||||
('team_name', models.CharField(blank=True, max_length=255, null=True)),
|
|
||||||
('group', models.CharField(max_length=255)),
|
|
||||||
('firstname', models.CharField(blank=True, max_length=255, null=True)),
|
|
||||||
('lastname', models.CharField(blank=True, max_length=255, null=True)),
|
|
||||||
('date_of_birth', models.DateField(blank=True, null=True)),
|
|
||||||
('female', models.BooleanField(default=False)),
|
|
||||||
('verification_code', models.UUIDField(default=uuid.uuid4, editable=False)),
|
|
||||||
('created_at', models.DateTimeField(auto_now_add=True)),
|
|
||||||
('expires_at', models.DateTimeField()),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.DeleteModel(
|
|
||||||
name='JpnAdminPerf',
|
|
||||||
),
|
|
||||||
migrations.DeleteModel(
|
|
||||||
name='JpnSubPerf',
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='date_of_birth',
|
|
||||||
field=models.DateField(blank=True, null=True),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='event_code',
|
|
||||||
field=models.CharField(blank=True, max_length=255, null=True, verbose_name='Event Code'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='female',
|
|
||||||
field=models.BooleanField(default=False),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='firstname',
|
|
||||||
field=models.CharField(blank=True, max_length=255, null=True),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='is_rogaining',
|
|
||||||
field=models.BooleanField(default=False),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='lastname',
|
|
||||||
field=models.CharField(blank=True, max_length=255, null=True),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='team_name',
|
|
||||||
field=models.CharField(blank=True, max_length=255, null=True, verbose_name='Team Name'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='zekken_number',
|
|
||||||
field=models.CharField(blank=True, max_length=255, null=True, verbose_name='Zekken Number'),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='email',
|
|
||||||
field=models.CharField(max_length=255, unique=True, verbose_name='Email'),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='location',
|
|
||||||
name='category',
|
|
||||||
field=models.CharField(blank=True, db_index=True, max_length=2048, null=True, verbose_name='Category'),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='location',
|
|
||||||
name='event_active',
|
|
||||||
field=models.BooleanField(db_index=True, default=True, verbose_name='Is Event active'),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='location',
|
|
||||||
name='event_name',
|
|
||||||
field=models.CharField(blank=True, db_index=True, max_length=2048, null=True, verbose_name='Event name'),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='location',
|
|
||||||
name='group',
|
|
||||||
field=models.CharField(blank=True, db_index=True, max_length=2048, null=True, verbose_name='Group'),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='location',
|
|
||||||
name='location_id',
|
|
||||||
field=models.IntegerField(blank=True, db_index=True, null=True, verbose_name='Location id'),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='templocation',
|
|
||||||
name='sub_loc_id',
|
|
||||||
field=models.CharField(blank=True, max_length=2048, null=True, verbose_name='Sub location id'),
|
|
||||||
),
|
|
||||||
migrations.AddIndex(
|
|
||||||
model_name='location',
|
|
||||||
index=django.contrib.postgres.indexes.GistIndex(fields=['geom'], name='rog_locatio_geom_4793cc_gist'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='team',
|
|
||||||
name='category',
|
|
||||||
field=models.ForeignKey(default=1, on_delete=django.db.models.deletion.CASCADE, to='rog.category'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='team',
|
|
||||||
name='owner',
|
|
||||||
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='owned_teams', to=settings.AUTH_USER_MODEL),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='member',
|
|
||||||
name='team',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.team'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='member',
|
|
||||||
name='user',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='goalimages',
|
|
||||||
name='user',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.DO_NOTHING, to=settings.AUTH_USER_MODEL),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='entrymember',
|
|
||||||
name='entry',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.entry'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='entrymember',
|
|
||||||
name='member',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.member'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='entry',
|
|
||||||
name='category',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.category'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='entry',
|
|
||||||
name='event',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.newevent'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='entry',
|
|
||||||
name='owner',
|
|
||||||
field=models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='entry',
|
|
||||||
name='team',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.team'),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='checkinimages',
|
|
||||||
name='user',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.DO_NOTHING, to=settings.AUTH_USER_MODEL),
|
|
||||||
),
|
|
||||||
migrations.AlterUniqueTogether(
|
|
||||||
name='category',
|
|
||||||
unique_together={('category_name', 'category_number')},
|
|
||||||
),
|
|
||||||
migrations.AlterUniqueTogether(
|
|
||||||
name='team',
|
|
||||||
unique_together={('zekken_number', 'category')},
|
|
||||||
),
|
|
||||||
migrations.AlterUniqueTogether(
|
|
||||||
name='member',
|
|
||||||
unique_together={('team', 'user')},
|
|
||||||
),
|
|
||||||
migrations.AlterUniqueTogether(
|
|
||||||
name='entrymember',
|
|
||||||
unique_together={('entry', 'member')},
|
|
||||||
),
|
|
||||||
migrations.AlterUniqueTogether(
|
|
||||||
name='entry',
|
|
||||||
unique_together={('team', 'event', 'date', 'owner')},
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,34 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-07-24 06:22
|
|
||||||
|
|
||||||
import datetime
|
|
||||||
from django.db import migrations, models
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0033_auto_20240724_1431'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='category',
|
|
||||||
name='category_number',
|
|
||||||
field=models.IntegerField(default=0),
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='NewCategory',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('category_name', models.CharField(max_length=255, unique=True)),
|
|
||||||
('category_number', models.IntegerField(default=0)),
|
|
||||||
('duration', models.DurationField(default=datetime.timedelta(seconds=18000))),
|
|
||||||
('num_of_member', models.IntegerField(default=1)),
|
|
||||||
('family', models.BooleanField(default=False)),
|
|
||||||
('female', models.BooleanField(default=False)),
|
|
||||||
],
|
|
||||||
options={
|
|
||||||
'unique_together': {('category_name', 'category_number')},
|
|
||||||
},
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,29 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-07-24 06:29
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
import django.db.models.deletion
|
|
||||||
import django.utils.timezone
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0034_auto_20240724_1522'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='NewEvent2',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('event_name', models.CharField(max_length=255, unique=True)),
|
|
||||||
('start_datetime', models.DateTimeField(default=django.utils.timezone.now)),
|
|
||||||
('end_datetime', models.DateTimeField()),
|
|
||||||
],
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='entry',
|
|
||||||
name='event',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.newevent2'),
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,20 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-07-24 06:58
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
import django.db.models.deletion
|
|
||||||
import rog.models
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0035_auto_20240724_1529'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='team',
|
|
||||||
name='category',
|
|
||||||
field=models.ForeignKey(default=rog.models.get_default_category, on_delete=django.db.models.deletion.SET_DEFAULT, to='rog.newcategory'),
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,19 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-07-24 15:20
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
import django.db.models.deletion
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0036_alter_team_category'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='member',
|
|
||||||
name='team',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='members', to='rog.team'),
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,19 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-07-25 01:21
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
import django.db.models.deletion
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0037_alter_member_team'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='entry',
|
|
||||||
name='category',
|
|
||||||
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='rog.newcategory'),
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,29 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-07-26 06:08
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
import django.utils.timezone
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0038_alter_entry_category'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='date_joined',
|
|
||||||
field=models.DateTimeField(default=django.utils.timezone.now),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='email',
|
|
||||||
field=models.EmailField(max_length=254, unique=True),
|
|
||||||
),
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name='customuser',
|
|
||||||
name='group',
|
|
||||||
field=models.CharField(blank=True, max_length=255),
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,33 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-08-01 08:29
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0039_auto_20240726_1508'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='member',
|
|
||||||
name='date_of_birth',
|
|
||||||
field=models.DateField(blank=True, null=True),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='member',
|
|
||||||
name='female',
|
|
||||||
field=models.BooleanField(default=False),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='member',
|
|
||||||
name='firstname',
|
|
||||||
field=models.CharField(blank=True, max_length=255, null=True),
|
|
||||||
),
|
|
||||||
migrations.AddField(
|
|
||||||
model_name='member',
|
|
||||||
name='lastname',
|
|
||||||
field=models.CharField(blank=True, max_length=255, null=True),
|
|
||||||
),
|
|
||||||
]
|
|
||||||
@ -1,37 +0,0 @@
|
|||||||
# Generated by Django 3.2.9 on 2024-08-02 15:11
|
|
||||||
|
|
||||||
import django.contrib.gis.db.models.fields
|
|
||||||
from django.db import migrations, models
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
('rog', '0040_auto_20240801_1729'),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.CreateModel(
|
|
||||||
name='JpnSubPerf',
|
|
||||||
fields=[
|
|
||||||
('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
|
|
||||||
('geom', django.contrib.gis.db.models.fields.MultiPolygonField(blank=True, null=True, srid=4326)),
|
|
||||||
('adm0_en', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm0_ja', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm0_pcode', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm1_en', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm1_ja', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm1_pcode', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm2_ja', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm2_en', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('adm2_pcode', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('name_modified', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('area_name', models.CharField(blank=True, max_length=254, null=True)),
|
|
||||||
('list_order', models.IntegerField(default=0)),
|
|
||||||
],
|
|
||||||
options={
|
|
||||||
'db_table': 'jpn_sub_perf',
|
|
||||||
'managed': False,
|
|
||||||
},
|
|
||||||
),
|
|
||||||
]
|
|
||||||
828
rog/models.py
828
rog/models.py
File diff suppressed because it is too large
Load Diff
@ -1,25 +0,0 @@
|
|||||||
|
|
||||||
from rest_framework import permissions
|
|
||||||
from .models import Team,Member
|
|
||||||
|
|
||||||
class IsMemberOrTeamOwner(permissions.BasePermission):
|
|
||||||
def has_object_permission(self, request, view, obj):
|
|
||||||
# Check if user is a member of the team or the team owner
|
|
||||||
return request.user in obj.team.members.all() or request.user == obj.team.owner
|
|
||||||
|
|
||||||
class IsTeamOwner(permissions.BasePermission):
|
|
||||||
def has_object_permission(self, request, view, obj):
|
|
||||||
if isinstance(obj, Team):
|
|
||||||
return obj.owner == request.user
|
|
||||||
elif isinstance(obj, Member):
|
|
||||||
return obj.team.owner == request.user
|
|
||||||
return False
|
|
||||||
|
|
||||||
class IsTeamOwnerOrMember(permissions.BasePermission):
|
|
||||||
def has_permission(self, request, view):
|
|
||||||
team_id = view.kwargs.get('team_id')
|
|
||||||
if not team_id:
|
|
||||||
return False
|
|
||||||
team = Team.objects.get(id=team_id)
|
|
||||||
return team.owner == request.user or team.members.filter(user=request.user).exists()
|
|
||||||
|
|
||||||
@ -1,342 +0,0 @@
|
|||||||
-- まず既存のビューをすべて削除
|
|
||||||
DROP MATERIALIZED VIEW IF EXISTS mv_entry_details CASCADE;
|
|
||||||
DROP VIEW IF EXISTS v_category_rankings CASCADE;
|
|
||||||
DROP VIEW IF EXISTS v_checkin_summary CASCADE;
|
|
||||||
|
|
||||||
-- チェックポイントの集計用ビュー
|
|
||||||
CREATE VIEW v_checkin_summary AS
|
|
||||||
SELECT
|
|
||||||
event_id,
|
|
||||||
event_code,
|
|
||||||
zekken_number, -- 文字列として保持
|
|
||||||
COUNT(*) as total_checkins,
|
|
||||||
COUNT(CASE WHEN buy_flag THEN 1 END) as purchase_count,
|
|
||||||
SUM(points) as total_points,
|
|
||||||
SUM(CASE WHEN buy_flag THEN points ELSE 0 END) as bonus_points,
|
|
||||||
SUM(CASE WHEN NOT buy_flag THEN points ELSE 0 END) as normal_points,
|
|
||||||
SUM(COALESCE(late_point, 0)) as penalty_points,
|
|
||||||
MAX(create_at) as last_checkin
|
|
||||||
FROM
|
|
||||||
gps_checkins
|
|
||||||
GROUP BY
|
|
||||||
event_id,event_code, zekken_number;
|
|
||||||
|
|
||||||
-- カテゴリー内ランキング計算用ビュー
|
|
||||||
CREATE VIEW v_category_rankings_old AS
|
|
||||||
SELECT
|
|
||||||
e.id,
|
|
||||||
e.event_id,
|
|
||||||
ev.event_name,
|
|
||||||
e.category_id,
|
|
||||||
CAST(e.zekken_number AS TEXT) as zekken_number, -- 数値を文字列に変換
|
|
||||||
COALESCE(cs.total_points, 0) as total_score,
|
|
||||||
RANK() OVER (PARTITION BY e.event_id, e.category_id
|
|
||||||
ORDER BY COALESCE(cs.total_points, 0) DESC) as ranking,
|
|
||||||
COUNT(*) OVER (PARTITION BY e.event_id, e.category_id) as total_participants
|
|
||||||
FROM
|
|
||||||
rog_entry e
|
|
||||||
JOIN rog_newevent2 ev ON e.event_id = ev.id
|
|
||||||
LEFT JOIN v_checkin_summary cs ON ev.event_name = cs.event_code
|
|
||||||
AND CAST(e.zekken_number AS TEXT) = cs.zekken_number
|
|
||||||
WHERE
|
|
||||||
e.is_active = true;
|
|
||||||
|
|
||||||
-- 完走状態を含むカテゴリーランキングビューの作成
|
|
||||||
-- 完走状態を含むカテゴリーランキングビューの作成
|
|
||||||
CREATE OR REPLACE VIEW v_category_rankings AS
|
|
||||||
WITH completion_status AS (
|
|
||||||
SELECT
|
|
||||||
e.id,
|
|
||||||
e.event_id,
|
|
||||||
e.category_id,
|
|
||||||
CAST(e.zekken_number AS TEXT) as zekken_number,
|
|
||||||
CASE
|
|
||||||
WHEN gi.goaltime IS NULL THEN '棄権'
|
|
||||||
WHEN gi.goaltime <= ev.end_datetime THEN '完走'
|
|
||||||
WHEN gi.goaltime > ev.end_datetime AND
|
|
||||||
gi.goaltime <= ev.end_datetime + INTERVAL '15 minutes' THEN '完走(遅刻)'
|
|
||||||
ELSE '失格'
|
|
||||||
END as completion_status,
|
|
||||||
COALESCE(cs.total_points, 0) as raw_points,
|
|
||||||
COALESCE(cs.normal_points, 0) as normal_points,
|
|
||||||
COALESCE(cs.bonus_points, 0) as bonus_points,
|
|
||||||
COALESCE(cs.penalty_points, 0) as original_penalty_points,
|
|
||||||
-- 遅刻ペナルティの計算(1秒でも遅れたら、その分数に応じて-50点/分)
|
|
||||||
CASE
|
|
||||||
WHEN gi.goaltime > ev.end_datetime THEN
|
|
||||||
(CEIL(EXTRACT(EPOCH FROM (gi.goaltime - ev.end_datetime)) / 60)) * (-50)
|
|
||||||
ELSE 0
|
|
||||||
END as late_penalty_points,
|
|
||||||
gi.goaltime,
|
|
||||||
ev.end_datetime
|
|
||||||
FROM
|
|
||||||
rog_entry e
|
|
||||||
JOIN rog_newevent2 ev ON e.event_id = ev.id
|
|
||||||
LEFT JOIN v_checkin_summary cs ON ev.event_name = cs.event_code
|
|
||||||
AND CAST(e.zekken_number AS TEXT) = cs.zekken_number
|
|
||||||
LEFT JOIN rog_goalimages gi ON e.owner_id = gi.user_id
|
|
||||||
AND gi.event_code = ev.event_name
|
|
||||||
WHERE
|
|
||||||
e.is_active = true
|
|
||||||
),
|
|
||||||
points_calculation AS (
|
|
||||||
SELECT
|
|
||||||
*,
|
|
||||||
-- 総合ポイントの再計算(遅刻ペナルティを含む)
|
|
||||||
raw_points + late_penalty_points as total_points
|
|
||||||
FROM completion_status
|
|
||||||
),
|
|
||||||
valid_rankings AS (
|
|
||||||
-- 完走者のみを対象とした順位付け
|
|
||||||
SELECT
|
|
||||||
*,
|
|
||||||
DENSE_RANK() OVER (
|
|
||||||
PARTITION BY event_id, category_id
|
|
||||||
ORDER BY
|
|
||||||
total_points DESC,
|
|
||||||
CASE
|
|
||||||
WHEN completion_status = '完走' THEN 1
|
|
||||||
WHEN completion_status = '完走(遅刻)' THEN 2
|
|
||||||
END,
|
|
||||||
goaltime
|
|
||||||
) as valid_rank
|
|
||||||
FROM points_calculation
|
|
||||||
WHERE completion_status IN ('完走', '完走(遅刻)')
|
|
||||||
)
|
|
||||||
SELECT
|
|
||||||
cs.id,
|
|
||||||
cs.event_id,
|
|
||||||
cs.category_id,
|
|
||||||
cs.zekken_number,
|
|
||||||
cs.raw_points as original_total_points,
|
|
||||||
cs.normal_points,
|
|
||||||
cs.bonus_points,
|
|
||||||
cs.original_penalty_points,
|
|
||||||
CASE
|
|
||||||
WHEN cs.completion_status IN ('完走(遅刻)', '失格') AND cs.goaltime IS NOT NULL THEN cs.late_penalty_points
|
|
||||||
ELSE 0
|
|
||||||
END as late_penalty_points,
|
|
||||||
ROUND(pc.total_points) as total_points,
|
|
||||||
cs.completion_status,
|
|
||||||
CASE
|
|
||||||
WHEN cs.completion_status IN ('完走', '完走(遅刻)') THEN CAST(vr.valid_rank AS TEXT)
|
|
||||||
WHEN cs.completion_status = '失格' THEN '失格'
|
|
||||||
WHEN cs.completion_status = '棄権' THEN '棄権'
|
|
||||||
END as ranking,
|
|
||||||
COUNT(*) FILTER (WHERE cs.completion_status IN ('完走', '完走(遅刻)'))
|
|
||||||
OVER (PARTITION BY cs.event_id, cs.category_id) as total_valid_participants
|
|
||||||
FROM
|
|
||||||
completion_status cs
|
|
||||||
JOIN points_calculation pc ON cs.id = pc.id
|
|
||||||
LEFT JOIN valid_rankings vr ON cs.id = vr.id;
|
|
||||||
|
|
||||||
|
|
||||||
-- マテリアライズドビューの作成
|
|
||||||
CREATE MATERIALIZED VIEW mv_entry_details AS
|
|
||||||
SELECT
|
|
||||||
-- 既存のフィールド
|
|
||||||
e.id,
|
|
||||||
CAST(e.zekken_number AS TEXT) as zekken_number,
|
|
||||||
e.is_active,
|
|
||||||
e."hasParticipated",
|
|
||||||
e."hasGoaled",
|
|
||||||
e.date as entry_date,
|
|
||||||
|
|
||||||
-- イベント情報
|
|
||||||
ev.event_name,
|
|
||||||
ev.event_description,
|
|
||||||
ev.start_datetime,
|
|
||||||
ev.end_datetime,
|
|
||||||
ev."deadlineDateTime",
|
|
||||||
TO_CHAR(ev.start_datetime::date, 'YYYY/MM/DD') as event_date,
|
|
||||||
|
|
||||||
-- カテゴリー情報
|
|
||||||
nc.category_name,
|
|
||||||
nc.category_number,
|
|
||||||
nc.duration,
|
|
||||||
nc.num_of_member,
|
|
||||||
nc.family as is_family_category,
|
|
||||||
nc.female as is_female_category,
|
|
||||||
|
|
||||||
-- チーム情報
|
|
||||||
t.team_name,
|
|
||||||
|
|
||||||
-- オーナー情報
|
|
||||||
cu.email as owner_email,
|
|
||||||
cu.firstname as owner_firstname,
|
|
||||||
cu.lastname as owner_lastname,
|
|
||||||
cu.date_of_birth as owner_birth_date,
|
|
||||||
cu.female as owner_is_female,
|
|
||||||
|
|
||||||
-- スコア情報
|
|
||||||
COALESCE(cs.normal_points, 0) as normal_points,
|
|
||||||
COALESCE(cs.bonus_points, 0) as bonus_points,
|
|
||||||
COALESCE(cs.total_checkins, 0) as checkin_count,
|
|
||||||
COALESCE(cs.purchase_count, 0) as purchase_count,
|
|
||||||
cr.late_penalty_points as penalty_points, -- 遅刻ペナルティを使用
|
|
||||||
cr.total_points as total_points, -- v_category_rankingsの総合ポイントを使用
|
|
||||||
|
|
||||||
|
|
||||||
-- ゴール情報
|
|
||||||
gi.goalimage as goal_image,
|
|
||||||
gi.goaltime as goal_time,
|
|
||||||
|
|
||||||
-- 完走状態の判定を追加
|
|
||||||
CASE
|
|
||||||
WHEN gi.goaltime IS NULL THEN '棄権'
|
|
||||||
WHEN gi.goaltime <= ev.end_datetime THEN '完走'
|
|
||||||
WHEN gi.goaltime > ev.end_datetime AND
|
|
||||||
gi.goaltime <= ev.end_datetime + INTERVAL '15 minutes' THEN '完走(遅刻)'
|
|
||||||
ELSE '失格'
|
|
||||||
END as validation,
|
|
||||||
|
|
||||||
-- ランキング情報
|
|
||||||
cr.ranking as category_rank,
|
|
||||||
cr.total_valid_participants,
|
|
||||||
|
|
||||||
-- チームメンバー情報(JSON形式で格納)
|
|
||||||
jsonb_agg(
|
|
||||||
jsonb_build_object(
|
|
||||||
'email', m.user_id,
|
|
||||||
'firstname', m.firstname,
|
|
||||||
'lastname', m.lastname,
|
|
||||||
'birth_date', m.date_of_birth,
|
|
||||||
'is_female', m.female,
|
|
||||||
'is_temporary', m.is_temporary,
|
|
||||||
'status', CASE
|
|
||||||
WHEN m.is_temporary THEN 'TEMPORARY'
|
|
||||||
WHEN m.date_of_birth IS NULL THEN 'PENDING'
|
|
||||||
ELSE 'ACTIVE'
|
|
||||||
END,
|
|
||||||
'member_type', CASE
|
|
||||||
WHEN m.user_id = e.owner_id THEN 'OWNER'
|
|
||||||
ELSE 'MEMBER'
|
|
||||||
END
|
|
||||||
) ORDER BY
|
|
||||||
CASE WHEN m.user_id = e.owner_id THEN 0 ELSE 1 END, -- オーナーを最初に
|
|
||||||
m.id
|
|
||||||
) FILTER (WHERE m.id IS NOT NULL) as team_members
|
|
||||||
|
|
||||||
FROM
|
|
||||||
rog_entry e
|
|
||||||
INNER JOIN rog_newevent2 ev ON e.event_id = ev.id
|
|
||||||
INNER JOIN rog_newcategory nc ON e.category_id = nc.id
|
|
||||||
INNER JOIN rog_team t ON e.team_id = t.id
|
|
||||||
LEFT JOIN rog_customuser cu ON e.owner_id = cu.id
|
|
||||||
LEFT JOIN v_checkin_summary cs ON e.event_id = cs.event_id -- この行を変更
|
|
||||||
AND CAST(e.zekken_number AS TEXT) = cs.zekken_number
|
|
||||||
LEFT JOIN v_category_rankings cr ON e.id = cr.id
|
|
||||||
LEFT JOIN rog_member m ON t.id = m.team_id
|
|
||||||
LEFT JOIN rog_goalimages gi ON e.owner_id = gi.user_id
|
|
||||||
AND gi.event_code = ev.event_name -- ゴール情報の結合条件も修正
|
|
||||||
|
|
||||||
GROUP BY
|
|
||||||
e.id, e.zekken_number, e.is_active, e."hasParticipated", e."hasGoaled", e.date,
|
|
||||||
ev.event_name,ev.event_description, ev.start_datetime, ev.end_datetime, ev."deadlineDateTime",
|
|
||||||
nc.category_name, nc.category_number, nc.duration, nc.num_of_member,
|
|
||||||
nc.family, nc.female,
|
|
||||||
t.team_name,
|
|
||||||
cu.email, cu.firstname, cu.lastname, cu.date_of_birth, cu.female,
|
|
||||||
cs.total_points, cs.normal_points, cs.bonus_points, cs.penalty_points,
|
|
||||||
cs.total_checkins, cs.purchase_count, cs.last_checkin,
|
|
||||||
cr.original_total_points, cr.late_penalty_points, cr.total_points,
|
|
||||||
cr.completion_status, cr.ranking, cr.total_valid_participants,
|
|
||||||
gi.goalimage, gi.goaltime,
|
|
||||||
e.owner_id;
|
|
||||||
|
|
||||||
-- インデックスの再作成
|
|
||||||
CREATE UNIQUE INDEX idx_mv_entry_details_event_zekken
|
|
||||||
ON mv_entry_details(id, event_name, zekken_number);
|
|
||||||
|
|
||||||
-- ビューの更新
|
|
||||||
REFRESH MATERIALIZED VIEW mv_entry_details;
|
|
||||||
|
|
||||||
|
|
||||||
-- チェックインと位置情報を結合したビューを作成
|
|
||||||
DROP VIEW IF EXISTS v_checkins_locations CASCADE;
|
|
||||||
CREATE OR REPLACE VIEW v_checkins_locations AS
|
|
||||||
SELECT
|
|
||||||
g.event_code,
|
|
||||||
g.zekken_number,
|
|
||||||
g.path_order,
|
|
||||||
g.cp_number,
|
|
||||||
l.sub_loc_id,
|
|
||||||
l.location_name,
|
|
||||||
l.photos,
|
|
||||||
g.image_address,
|
|
||||||
g.create_at,
|
|
||||||
g.buy_flag,
|
|
||||||
g.validate_location,
|
|
||||||
g.points
|
|
||||||
FROM
|
|
||||||
gps_checkins g
|
|
||||||
LEFT JOIN rog_location l ON g.cp_number = l.cp
|
|
||||||
AND l."group" LIKE '%' || g.event_code || '%'
|
|
||||||
ORDER BY
|
|
||||||
g.event_code,
|
|
||||||
g.zekken_number,
|
|
||||||
g.path_order;
|
|
||||||
|
|
||||||
-- インデックスのサジェスチョン(実際のテーブルに適用する必要があります)
|
|
||||||
/*
|
|
||||||
CREATE INDEX idx_gps_checkins_cp_number ON gps_checkins(cp_number);
|
|
||||||
CREATE INDEX idx_rog_location_cp ON rog_location(cp);
|
|
||||||
*/
|
|
||||||
|
|
||||||
-- チェックポイントごとの集計VIEW
|
|
||||||
-- チェックポイントごとの集計ビューを作成
|
|
||||||
DROP VIEW IF EXISTS v_checkpoint_summary CASCADE;
|
|
||||||
CREATE OR REPLACE VIEW v_checkpoint_summary AS
|
|
||||||
WITH checkpoint_counts AS (
|
|
||||||
SELECT
|
|
||||||
e.event_id,
|
|
||||||
ev.event_name,
|
|
||||||
gc.cp_number,
|
|
||||||
l.sub_loc_id,
|
|
||||||
l.location_name,
|
|
||||||
e.category_id,
|
|
||||||
nc.category_name,
|
|
||||||
COUNT(CASE
|
|
||||||
WHEN gc.validate_location = true AND gc.buy_flag = false
|
|
||||||
THEN 1
|
|
||||||
END) as normal_checkins,
|
|
||||||
COUNT(CASE
|
|
||||||
WHEN gc.validate_location = true AND gc.buy_flag = true
|
|
||||||
THEN 1
|
|
||||||
END) as purchase_checkins
|
|
||||||
FROM
|
|
||||||
rog_entry e
|
|
||||||
JOIN rog_newevent2 ev ON e.event_id = ev.id
|
|
||||||
JOIN rog_newcategory nc ON e.category_id = nc.id
|
|
||||||
JOIN gps_checkins gc ON ev.event_name = gc.event_code
|
|
||||||
AND CAST(e.zekken_number AS TEXT) = gc.zekken_number
|
|
||||||
LEFT JOIN rog_location l ON gc.cp_number = l.cp
|
|
||||||
AND l."group" LIKE '%' || gc.event_code || '%'
|
|
||||||
WHERE
|
|
||||||
e.is_active = true
|
|
||||||
AND gc.validate_location = true
|
|
||||||
GROUP BY
|
|
||||||
e.event_id,
|
|
||||||
ev.event_name,
|
|
||||||
gc.cp_number,
|
|
||||||
l.sub_loc_id,
|
|
||||||
l.location_name,
|
|
||||||
e.category_id,
|
|
||||||
nc.category_name
|
|
||||||
)
|
|
||||||
SELECT
|
|
||||||
event_id,
|
|
||||||
event_name,
|
|
||||||
cp_number,
|
|
||||||
sub_loc_id,
|
|
||||||
location_name,
|
|
||||||
category_id,
|
|
||||||
category_name,
|
|
||||||
normal_checkins,
|
|
||||||
purchase_checkins
|
|
||||||
FROM
|
|
||||||
checkpoint_counts
|
|
||||||
ORDER BY
|
|
||||||
event_name,
|
|
||||||
cp_number,
|
|
||||||
category_id;
|
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user