mirror of
https://github.com/Ladebeze66/AIagent.git
synced 2025-12-13 09:06:49 +01:00
J9
This commit is contained in:
parent
7ab516961f
commit
5128ea8904
256
JSON/messages.json
Normal file
256
JSON/messages.json
Normal file
@ -0,0 +1,256 @@
|
||||
[
|
||||
{
|
||||
"id": 227288,
|
||||
"body": "",
|
||||
"date": "2025-03-12 08:14:38",
|
||||
"author_id": [
|
||||
28961,
|
||||
"Fabien LAFAY"
|
||||
],
|
||||
"email_from": "\"Fabien LAFAY\" <fabien@cbao.fr>",
|
||||
"subject": false,
|
||||
"parent_id": false,
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
2,
|
||||
"Note"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225765,
|
||||
"body": "Bonjour , Je tiens à m'excuser pour ce désagrément. J'ai relancé Quentin, qui s'engage à le mettre en ligne au plus vite. Je garde ce ticket ouvert jusqu'à sa mise à disposition afin de suivre votre demande de près. Je reste à votre entière disposition pour toute information complémentaire. Cordialement, --- Support technique Afin d'assurer une meilleure traçabilité et vous garantir une prise en charge optimale, nous vous invitons à envoyer vos demandes d'assistance technique à support@cbao.fr L'objectif du Support Technique est de vous aider : si vous rencontrez une difficulté, ou pour nous soumettre une ou des suggestions d'amélioration de nos logiciels ou de nos méthodes. Notre service est ouvert du lundi au vendredi de 9h à 12h et de 14h à 18h. Dès réception, un technicien prendra en charge votre demande et au besoin vous rappellera. Confidentialité : Ce courriel contient des informations confidentielles exclusivement réservées au destinataire mentionné. Si vous deviez recevoir cet e-mail par erreur, merci d’en avertir immédiatement l’expéditeur et de le supprimer de votre système informatique. Au cas où vous ne seriez pas destinataire de ce message, veuillez noter que sa divulgation, sa copie ou tout acte en rapport avec la communication du contenu des informations est strictement interdit.",
|
||||
"date": "2025-02-21 15:51:53",
|
||||
"author_id": [
|
||||
32165,
|
||||
"Romuald GRUSON"
|
||||
],
|
||||
"email_from": "support@cbao.fr",
|
||||
"subject": "Re: [T10929] - TR: MAJ BRGlab",
|
||||
"parent_id": [
|
||||
225611,
|
||||
"[T10929] TR: MAJ BRGlab"
|
||||
],
|
||||
"message_type": "comment",
|
||||
"subtype_id": [
|
||||
1,
|
||||
"Discussions"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225756,
|
||||
"body": "",
|
||||
"date": "2025-02-21 15:17:55",
|
||||
"author_id": [
|
||||
32165,
|
||||
"Romuald GRUSON"
|
||||
],
|
||||
"email_from": "\"Romuald GRUSON\" <romuald@cbao.fr>",
|
||||
"subject": false,
|
||||
"parent_id": [
|
||||
225611,
|
||||
"[T10929] TR: MAJ BRGlab"
|
||||
],
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
16,
|
||||
"Task Created"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225755,
|
||||
"body": "Bonjour, L'essai présent dans notre base de données est une tentative d'essai réalisée par moi qui ne fonctionne pas. En cause le module ZK qui ne gère pas efficacement l'affichage des résultats en puissance de 10. J'avais fait un ticket en ce sens il y a 2 ans et demi à ce sujet. Le ticket avait été marqué comme étant une priorité il y a déjà plusieurs mois, et Quentin devait le passer en mode système en le créant en code. Lors de notre entretien sur Teams le 05/02, Quentin m'avait informé que l'essai serait disponible lors de la mise à jour prévue le lundi 10/02. Pouvez-vous me donner plus d'informations au sujet de cet essai ? Je tiens à préciser qu'après 2 ans et demi d'attente, le critère d'urgence peut paraitre relatif, mais à ce jour nous en avons un besoin impératif. Cordialement, Olivier Antoni Technicien Egis Géotechnique ( 07 88 25 39 98 olivier.antoni@egis -group.com I www.egis.fr Egis Géotechnique 3 rue docteur Schweitzer 38180 Seyssins FRANCE Suivez Egis sur : P Afin de contribuer au respect de l'environnement, merci de n'imprimer ce mail qu'en cas de nécessité De : support@cbao.fr <support@cbao.fr> Envoyé : jeudi 20 février 2025 14:13 À : ANTONI Olivier <Olivier.ANTONI@egis-group.com> Objet : Re: [T10929] - TR: MAJ BRGlab /!\\ Courriel externe - Merci d'être prudent avec les liens et les pièces jointes /!\\ External email - Please be careful with links and attachments /!\\ Bonjour , Je constate que la norme est déjà présente dans votre base de données. Est-ce sur un type de matériau spécifique que vous ne pouvez pas y accéder ? Pouvez-vous préciser le contexte afin que nous puissions identifier l'origine du problème ? Je reste à votre entière disposition pour toute information complémentaire. Cordialement, --- Support technique Afin d'assurer une meilleure traçabilité et vous garantir une prise en charge optimale, nous vous invitons à envoyer vos demandes d'assistance technique à support@cbao.fr L'objectif du Support Technique est de vous aider : si vous rencontrez une difficulté, ou pour nous soumettre une ou des suggestions d'amélioration de nos logiciels ou de nos méthodes. Notre service est ouvert du lundi au vendredi de 9h à 12h et de 14h à 18h. Dès réception, un technicien prendra en charge votre demande et au besoin vous rappellera. Confidentialité : Ce courriel contient des informations confidentielles exclusivement réservées au destinataire mentionné. Si vous deviez recevoir cet e-mail par erreur, merci d’en avertir immédiatement l’expéditeur et de le supprimer de votre système informatique. Au cas où vous ne seriez pas destinataire de ce message, veuillez noter que sa divulgation, sa copie ou tout acte en rapport avec la communication du contenu des informations est strictement interdit. Envoyé par CBAO S.A.R.L. . Ce message et ses pièces jointes peuvent contenir des informations confidentielles ou privilégiées et ne doivent donc pas être diffusés, exploités ou copiés sans autorisation. Si vous avez reçu ce message par erreur, merci de le signaler à l'expéditeur et le détruire ainsi que les pièces jointes. Les messages électroniques étant susceptibles d'altération, Egis décline toute responsabilité si ce message a été altéré, déformé ou falsifié. Merci. This message and its attachments may contain confidential or privileged information that may be protected by law; they should not be distributed, used or copied without authorisation. If you have received this email in error, please notify the sender and delete this message and its attachments. As emails may be altered, Egis is not liable for messages that have been modified, changed or falsified. Thank you.",
|
||||
"date": "2025-02-21 15:14:30",
|
||||
"author_id": [
|
||||
5654,
|
||||
"EGIS SA, Olivier ANTONI"
|
||||
],
|
||||
"email_from": "ANTONI Olivier <Olivier.ANTONI@egis-group.com>",
|
||||
"subject": "RE: [T10929] - TR: MAJ BRGlab",
|
||||
"parent_id": [
|
||||
225611,
|
||||
"[T10929] TR: MAJ BRGlab"
|
||||
],
|
||||
"message_type": "email",
|
||||
"subtype_id": [
|
||||
1,
|
||||
"Discussions"
|
||||
],
|
||||
"attachment_ids": [
|
||||
143501,
|
||||
143499,
|
||||
143497,
|
||||
143495,
|
||||
143493,
|
||||
143491
|
||||
]
|
||||
},
|
||||
{
|
||||
"id": 225636,
|
||||
"body": "",
|
||||
"date": "2025-02-20 13:13:34",
|
||||
"author_id": [
|
||||
32165,
|
||||
"Romuald GRUSON"
|
||||
],
|
||||
"email_from": "\"Romuald GRUSON\" <romuald@cbao.fr>",
|
||||
"subject": false,
|
||||
"parent_id": [
|
||||
225611,
|
||||
"[T10929] TR: MAJ BRGlab"
|
||||
],
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
19,
|
||||
"Stage Changed"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225635,
|
||||
"body": "",
|
||||
"date": "2025-02-20 13:13:34",
|
||||
"author_id": [
|
||||
32165,
|
||||
"Romuald GRUSON"
|
||||
],
|
||||
"email_from": "\"Romuald GRUSON\" <romuald@cbao.fr>",
|
||||
"subject": false,
|
||||
"parent_id": false,
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
2,
|
||||
"Note"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225634,
|
||||
"body": "",
|
||||
"date": "2025-02-20 13:13:32",
|
||||
"author_id": [
|
||||
32165,
|
||||
"Romuald GRUSON"
|
||||
],
|
||||
"email_from": "\"Romuald GRUSON\" <romuald@cbao.fr>",
|
||||
"subject": false,
|
||||
"parent_id": false,
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
2,
|
||||
"Note"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225633,
|
||||
"body": "Bonjour , Je constate que la norme est déjà présente dans votre base de données. Est-ce sur un type de matériau spécifique que vous ne pouvez pas y accéder ? Pouvez-vous préciser le contexte afin que nous puissions identifier l'origine du problème ? Je reste à votre entière disposition pour toute information complémentaire. Cordialement, --- Support technique Afin d'assurer une meilleure traçabilité et vous garantir une prise en charge optimale, nous vous invitons à envoyer vos demandes d'assistance technique à support@cbao.fr L'objectif du Support Technique est de vous aider : si vous rencontrez une difficulté, ou pour nous soumettre une ou des suggestions d'amélioration de nos logiciels ou de nos méthodes. Notre service est ouvert du lundi au vendredi de 9h à 12h et de 14h à 18h. Dès réception, un technicien prendra en charge votre demande et au besoin vous rappellera. Confidentialité : Ce courriel contient des informations confidentielles exclusivement réservées au destinataire mentionné. Si vous deviez recevoir cet e-mail par erreur, merci d’en avertir immédiatement l’expéditeur et de le supprimer de votre système informatique. Au cas où vous ne seriez pas destinataire de ce message, veuillez noter que sa divulgation, sa copie ou tout acte en rapport avec la communication du contenu des informations est strictement interdit.",
|
||||
"date": "2025-02-20 13:13:30",
|
||||
"author_id": [
|
||||
32165,
|
||||
"Romuald GRUSON"
|
||||
],
|
||||
"email_from": "support@cbao.fr",
|
||||
"subject": "Re: [T10929] - TR: MAJ BRGlab",
|
||||
"parent_id": [
|
||||
225611,
|
||||
"[T10929] TR: MAJ BRGlab"
|
||||
],
|
||||
"message_type": "comment",
|
||||
"subtype_id": [
|
||||
1,
|
||||
"Discussions"
|
||||
],
|
||||
"attachment_ids": [
|
||||
143414
|
||||
]
|
||||
},
|
||||
{
|
||||
"id": 225614,
|
||||
"body": "",
|
||||
"date": "2025-02-20 09:37:17",
|
||||
"author_id": [
|
||||
2,
|
||||
"OdooBot"
|
||||
],
|
||||
"email_from": "\"OdooBot\" <odoobot@example.com>",
|
||||
"subject": false,
|
||||
"parent_id": false,
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
2,
|
||||
"Note"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225613,
|
||||
"body": "",
|
||||
"date": "2025-02-20 09:37:17",
|
||||
"author_id": [
|
||||
2,
|
||||
"OdooBot"
|
||||
],
|
||||
"email_from": "\"OdooBot\" <odoobot@example.com>",
|
||||
"subject": false,
|
||||
"parent_id": false,
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
2,
|
||||
"Note"
|
||||
],
|
||||
"attachment_ids": []
|
||||
},
|
||||
{
|
||||
"id": 225612,
|
||||
"body": "Bonjour Quentin, Je constate que l'essai de perméabilité n'a pas été ajouté, penses-tu qu'il puisse l'être rapidement ? Cordialement, Olivier Antoni Technicien Egis Géotechnique ( 07 88 25 39 98 olivier.antoni@egis -group.com I www.egis.fr Egis Géotechnique 3 rue docteur Schweitzer 38180 Seyssins FRANCE Suivez Egis sur : P Afin de contribuer au respect de l'environnement, merci de n'imprimer ce mail qu'en cas de nécessité De : ANTONI Olivier <Olivier.ANTONI@egis-group.com> Envoyé : lundi 17 février 2025 11:23 À : Quentin Faivre <quentin.faivre@cbao.fr> Objet : MAJ BRGlab Salut Quentin, Est-ce que tu as pu faire la mise à jour de BRGLab, avec notamment l'ajout de l'essai de perméabilité (NF EN ISO 17892-11) ? Cordialement, Olivier. Olivier Antoni Technicien Egis Géotechnique ( 07 88 25 39 98 olivier.antoni@egis -group.com I www.egis.fr Egis Géotechnique 3 rue docteur Schweitzer 38180 Seyssins FRANCE Suivez Egis sur : P Afin de contribuer au respect de l'environnement, merci de n'imprimer ce mail qu'en cas de nécessité Ce message et ses pièces jointes peuvent contenir des informations confidentielles ou privilégiées et ne doivent donc pas être diffusés, exploités ou copiés sans autorisation. Si vous avez reçu ce message par erreur, merci de le signaler à l'expéditeur et le détruire ainsi que les pièces jointes. Les messages électroniques étant susceptibles d'altération, Egis décline toute responsabilité si ce message a été altéré, déformé ou falsifié. Merci. This message and its attachments may contain confidential or privileged information that may be protected by law; they should not be distributed, used or copied without authorisation. If you have received this email in error, please notify the sender and delete this message and its attachments. As emails may be altered, Egis is not liable for messages that have been modified, changed or falsified. Thank you.",
|
||||
"date": "2025-02-20 09:36:33",
|
||||
"author_id": [
|
||||
5654,
|
||||
"EGIS SA, Olivier ANTONI"
|
||||
],
|
||||
"email_from": "ANTONI Olivier <Olivier.ANTONI@egis-group.com>",
|
||||
"subject": "TR: MAJ BRGlab",
|
||||
"parent_id": [
|
||||
225611,
|
||||
"[T10929] TR: MAJ BRGlab"
|
||||
],
|
||||
"message_type": "email",
|
||||
"subtype_id": [
|
||||
1,
|
||||
"Discussions"
|
||||
],
|
||||
"attachment_ids": [
|
||||
143378,
|
||||
143376,
|
||||
143374,
|
||||
143372,
|
||||
143370,
|
||||
143368,
|
||||
143366,
|
||||
143364,
|
||||
143362,
|
||||
143360
|
||||
]
|
||||
},
|
||||
{
|
||||
"id": 225611,
|
||||
"body": "",
|
||||
"date": "2025-02-20 09:37:16",
|
||||
"author_id": [
|
||||
2,
|
||||
"OdooBot"
|
||||
],
|
||||
"email_from": "\"OdooBot\" <odoobot@example.com>",
|
||||
"subject": false,
|
||||
"parent_id": false,
|
||||
"message_type": "notification",
|
||||
"subtype_id": [
|
||||
16,
|
||||
"Task Created"
|
||||
],
|
||||
"attachment_ids": []
|
||||
}
|
||||
]
|
||||
@ -1,444 +0,0 @@
|
||||
# Guide complet sur Ollama
|
||||
|
||||
## 1. Connaissances de base (commandes complètes, commandes NVIDIA pour H100)
|
||||
|
||||
### 1.1 Commandes essentielles d'Ollama
|
||||
|
||||
Voici les commandes fondamentales pour utiliser Ollama efficacement :
|
||||
|
||||
| Commande | Description |
|
||||
|----------|-------------|
|
||||
| `ollama serve` | Démarre le serveur Ollama sur votre système local |
|
||||
| `ollama create <nouveau_modèle>` | Crée un nouveau modèle à partir d'un modèle existant |
|
||||
| `ollama show <modèle>` | Affiche les détails d'un modèle spécifique |
|
||||
| `ollama run <modèle>` | Exécute le modèle spécifié pour interaction |
|
||||
| `ollama pull <modèle>` | Télécharge le modèle spécifié sur votre système |
|
||||
| `ollama list` | Liste tous les modèles téléchargés |
|
||||
| `ollama ps` | Affiche les modèles actuellement en cours d'exécution |
|
||||
| `ollama stop <modèle>` | Arrête le modèle spécifié en cours d'exécution |
|
||||
| `ollama rm <modèle>` | Supprime le modèle spécifié de votre système |
|
||||
|
||||
### 1.2 Commandes pour l'utilisation avec GPU NVIDIA (H100)
|
||||
|
||||
Pour tirer parti des GPU NVIDIA H100 avec Ollama :
|
||||
|
||||
1. **Vérification de l'installation NVIDIA** :
|
||||
```bash
|
||||
nvidia-smi
|
||||
```
|
||||
|
||||
2. **Variables d'environnement pour l'exécution sur GPU** :
|
||||
```bash
|
||||
__NV_PRIME_RENDER_OFFLOAD=1 __GLX_VENDOR_LIBRARY_NAME=nvidia ollama run <modèle>
|
||||
```
|
||||
|
||||
3. **Vérification de l'utilisation GPU** :
|
||||
```bash
|
||||
__NV_PRIME_RENDER_OFFLOAD=1 __GLX_VENDOR_LIBRARY_NAME=nvidia glxinfo | grep vendor
|
||||
```
|
||||
|
||||
4. **Commandes en cas de problème après mise en veille** (Linux) :
|
||||
```bash
|
||||
sudo rmmod nvidia_uvm && sudo modprobe nvidia_uvm
|
||||
```
|
||||
|
||||
## 2. Connaissances globales de Ollama
|
||||
|
||||
### 2.1 Qu'est-ce qu'Ollama ?
|
||||
|
||||
Ollama est une plateforme qui permet d'exécuter des modèles de langage localement sur votre propre ordinateur. Cela facilite l'utilisation de l'IA pour les développeurs et les entreprises sans dépendre de serveurs externes ou d'une connexion internet. La plateforme simplifie le processus d'intégration de l'IA dans vos applications en fournissant des outils pour construire, entraîner et déployer des modèles directement depuis votre environnement local.
|
||||
|
||||
### 2.2 Modèles disponibles
|
||||
|
||||
Ollama prend en charge une large gamme de modèles. Voici quelques exemples :
|
||||
|
||||
| Modèle | Paramètres | Taille | Commande |
|
||||
|--------|------------|--------|----------|
|
||||
| Llama 3.1 | 8B | 4.7GB | `ollama run llama3.1` |
|
||||
| Llama 3.1 | 70B | 40GB | `ollama run llama3.1:70b` |
|
||||
| Phi 3 Mini | 3.8B | 2.3GB | `ollama run phi3` |
|
||||
| Phi 3 Medium | 14B | 7.9GB | `ollama run phi3:medium` |
|
||||
| Gemma 2 | 27B | 16GB | `ollama run gemma2:27b` |
|
||||
| Mistral | 7B | 4.1GB | `ollama run mistral` |
|
||||
| Code Llama | 7B | 3.8GB | `ollama run codellama` |
|
||||
|
||||
### 2.3 Configuration de base
|
||||
|
||||
La configuration d'Ollama se fait généralement dans le fichier `~/.ollama/config.json` :
|
||||
```json
|
||||
{
|
||||
"host": "127.0.0.1",
|
||||
"port": 11434,
|
||||
"timeout": 30
|
||||
}
|
||||
```
|
||||
|
||||
## 3. Connaissances poussées
|
||||
|
||||
### 3.1 Exécution dockerisée
|
||||
|
||||
Pour utiliser Ollama dans Docker avec support GPU :
|
||||
|
||||
```bash
|
||||
docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
|
||||
```
|
||||
|
||||
### 3.2 Création de modèles personnalisés
|
||||
|
||||
La création d'un modèle personnalisé se fait via un fichier Modelfile :
|
||||
|
||||
```
|
||||
FROM llama3.1
|
||||
SYSTEM Je suis un assistant spécialisé dans le développement logiciel.
|
||||
PARAMETER num_ctx 8192
|
||||
PARAMETER temperature 0.7
|
||||
PARAMETER num_gpu 1
|
||||
```
|
||||
|
||||
Puis créez le modèle :
|
||||
```bash
|
||||
ollama create mon-assistant -f ./Modelfile
|
||||
```
|
||||
|
||||
### 3.3 Configuration avancée pour les GPU H100
|
||||
|
||||
Les H100 sont les GPU les plus puissants de NVIDIA pour l'IA :
|
||||
|
||||
- Utilisent des cœurs tenseurs de 4ème génération
|
||||
- Supportent le FP8 pour une meilleure précision
|
||||
- Offrent une bande passante mémoire de 3 TB/s
|
||||
- Fournissent 60 téraflops en double précision (FP64)
|
||||
- Permettent de gérer des modèles très larges comme Falcon 180B
|
||||
|
||||
## 4. Approfondissement partie API
|
||||
|
||||
### 4.1 Utilisation de l'API avec curl
|
||||
|
||||
L'API Ollama est accessible via HTTP sur le port 11434 par défaut :
|
||||
|
||||
```bash
|
||||
# Génération simple
|
||||
curl http://localhost:11434/api/generate -d '{
|
||||
"model": "llama3.1",
|
||||
"prompt": "Comment vas-tu aujourd'hui?"
|
||||
}'
|
||||
|
||||
# Désactiver le streaming pour obtenir une réponse complète
|
||||
curl http://localhost:11434/api/generate -d '{
|
||||
"model": "llama3.1",
|
||||
"prompt": "Comment vas-tu aujourd'hui?",
|
||||
"stream": false
|
||||
}'
|
||||
```
|
||||
|
||||
### 4.2 Intégration avec Python
|
||||
|
||||
Installation :
|
||||
```bash
|
||||
pip install ollama
|
||||
```
|
||||
|
||||
Exemple d'utilisation :
|
||||
```python
|
||||
import ollama
|
||||
|
||||
# Génération simple
|
||||
response = ollama.chat(
|
||||
model="llama3.1",
|
||||
messages=[
|
||||
{"role": "user", "content": "Explique-moi les bases du machine learning."}
|
||||
]
|
||||
)
|
||||
print(response['message']['content'])
|
||||
|
||||
# Avec streaming
|
||||
response = ollama.chat(
|
||||
model="llama3.1",
|
||||
messages=[
|
||||
{"role": "user", "content": "Écris un poème sur l'IA."}
|
||||
],
|
||||
stream=True
|
||||
)
|
||||
|
||||
for chunk in response:
|
||||
print(chunk['message']['content'], end='', flush=True)
|
||||
```
|
||||
|
||||
### 4.3 Endpoints principaux de l'API
|
||||
|
||||
- `/api/generate` : Génère du texte à partir d'un prompt
|
||||
- `/api/chat` : Conversation avec historique
|
||||
- `/api/embeddings` : Génère des embeddings vectoriels
|
||||
- `/api/models` : Liste ou gère les modèles
|
||||
- `/api/pull` : Télécharge un modèle
|
||||
- `/api/create` : Crée un modèle personnalisé
|
||||
|
||||
## 5. Création d'agent avec Ollama
|
||||
|
||||
### 5.1 Structure de base d'un agent Ollama
|
||||
|
||||
Voici un exemple de classe d'agent Ollama en Python :
|
||||
|
||||
```python
|
||||
from typing import List, Dict, Optional, AsyncIterable, Any
|
||||
from dataclasses import dataclass
|
||||
|
||||
@dataclass
|
||||
class OllamaAgentOptions:
|
||||
streaming: bool = False
|
||||
model_id: str = "llama2"
|
||||
# Paramètres additionnels (temperature, top_k, top_p, etc.)
|
||||
|
||||
class OllamaAgent:
|
||||
def __init__(self, options: OllamaAgentOptions):
|
||||
self.model_id = options.model_id
|
||||
self.streaming = options.streaming
|
||||
# Initialisation d'autres attributs
|
||||
|
||||
async def process_request(
|
||||
self,
|
||||
input_text: str,
|
||||
user_id: str,
|
||||
session_id: str,
|
||||
chat_history: List[Dict],
|
||||
additional_params: Optional[Dict[str, str]] = None
|
||||
):
|
||||
messages = [
|
||||
{"role": msg["role"], "content": msg["content"]}
|
||||
for msg in chat_history
|
||||
]
|
||||
messages.append({"role": "user", "content": input_text})
|
||||
|
||||
if self.streaming:
|
||||
return await self.handle_streaming_response(messages)
|
||||
else:
|
||||
response = ollama.chat(
|
||||
model=self.model_id,
|
||||
messages=messages
|
||||
)
|
||||
return {
|
||||
"role": "assistant",
|
||||
"content": response['message']['content']
|
||||
}
|
||||
|
||||
async def handle_streaming_response(self, messages: List[Dict[str, str]]):
|
||||
text = ''
|
||||
try:
|
||||
response = ollama.chat(
|
||||
model=self.model_id,
|
||||
messages=messages,
|
||||
stream=True
|
||||
)
|
||||
for part in response:
|
||||
text += part['message']['content']
|
||||
yield part['message']['content']
|
||||
|
||||
return {
|
||||
"role": "assistant",
|
||||
"content": text
|
||||
}
|
||||
except Exception as error:
|
||||
print(f"Erreur lors du streaming : {error}")
|
||||
raise error
|
||||
```
|
||||
|
||||
### 5.2 Création d'un agent spécialisé
|
||||
|
||||
Exemple d'un agent spécialisé pour la programmation :
|
||||
|
||||
```python
|
||||
class OllamaCodingAgent(OllamaAgent):
|
||||
def __init__(self):
|
||||
super().__init__(OllamaAgentOptions(
|
||||
model_id="codellama",
|
||||
streaming=True
|
||||
))
|
||||
self.system_prompt = "Tu es un assistant de programmation expert. Réponds avec du code clair et bien structuré."
|
||||
|
||||
async def process_request(self, input_text, user_id, session_id, chat_history, additional_params=None):
|
||||
# Ajouter le system prompt au début de l'historique
|
||||
if not chat_history or chat_history[0]["role"] != "system":
|
||||
chat_history.insert(0, {"role": "system", "content": self.system_prompt})
|
||||
|
||||
return await super().process_request(input_text, user_id, session_id, chat_history, additional_params)
|
||||
```
|
||||
|
||||
### 5.3 Orchestration multi-agents
|
||||
|
||||
Exemple d'orchestration de plusieurs agents Ollama :
|
||||
|
||||
```python
|
||||
class OllamaOrchestrator:
|
||||
def __init__(self):
|
||||
self.agents = {}
|
||||
self.sessions = {}
|
||||
|
||||
def add_agent(self, agent_id, agent):
|
||||
self.agents[agent_id] = agent
|
||||
|
||||
def create_session(self, session_id, user_id):
|
||||
self.sessions[session_id] = {
|
||||
"user_id": user_id,
|
||||
"history": []
|
||||
}
|
||||
|
||||
async def route_request(self, session_id, agent_id, input_text):
|
||||
if session_id not in self.sessions:
|
||||
self.create_session(session_id, "default_user")
|
||||
|
||||
if agent_id not in self.agents:
|
||||
raise ValueError(f"Agent {agent_id} non trouvé")
|
||||
|
||||
agent = self.agents[agent_id]
|
||||
session = self.sessions[session_id]
|
||||
|
||||
response = await agent.process_request(
|
||||
input_text=input_text,
|
||||
user_id=session["user_id"],
|
||||
session_id=session_id,
|
||||
chat_history=session["history"]
|
||||
)
|
||||
|
||||
session["history"].append({"role": "user", "content": input_text})
|
||||
session["history"].append(response)
|
||||
|
||||
return response
|
||||
```
|
||||
|
||||
## 6. Optimisation rapide et pratique du paramétrage des modèles
|
||||
|
||||
### 6.1 Paramètres principaux à optimiser
|
||||
|
||||
Voici les paramètres clés pour optimiser les performances d'Ollama :
|
||||
|
||||
| Paramètre | Description | Valeurs recommandées |
|
||||
|-----------|-------------|---------------------|
|
||||
| `temperature` | Contrôle la créativité des réponses | 0.0-1.0 (0.7 par défaut) |
|
||||
| `top_p` | Sampling de probabilité cumulative | 0.1-1.0 (0.9 par défaut) |
|
||||
| `top_k` | Nombre de tokens à considérer | 1-100 (40 par défaut) |
|
||||
| `num_ctx` | Taille du contexte en tokens | 2048-8192 selon modèle |
|
||||
| `num_gpu` | Nombre de GPU à utiliser | 1 ou plus selon disponibilité |
|
||||
| `num_thread` | Nombre de threads CPU | Dépend de votre processeur |
|
||||
| `repeat_penalty` | Pénalité pour répétition | 1.0-1.3 (1.1 par défaut) |
|
||||
|
||||
### 6.2 Optimisation selon le cas d'usage
|
||||
|
||||
#### Pour la génération de code :
|
||||
```
|
||||
PARAMETER temperature 0.2
|
||||
PARAMETER top_p 0.95
|
||||
PARAMETER repeat_penalty 1.2
|
||||
PARAMETER num_ctx 8192
|
||||
```
|
||||
|
||||
#### Pour la créativité :
|
||||
```
|
||||
PARAMETER temperature 0.8
|
||||
PARAMETER top_p 0.9
|
||||
PARAMETER repeat_penalty 1.05
|
||||
PARAMETER num_ctx 4096
|
||||
```
|
||||
|
||||
#### Pour les réponses factuelles :
|
||||
```
|
||||
PARAMETER temperature 0.1
|
||||
PARAMETER top_p 0.8
|
||||
PARAMETER repeat_penalty 1.2
|
||||
PARAMETER num_ctx 4096
|
||||
```
|
||||
|
||||
### 6.3 Technique d'itération pour l'optimisation
|
||||
|
||||
1. Commencer avec les paramètres par défaut
|
||||
2. Ajuster un paramètre à la fois et évaluer les résultats
|
||||
3. Documenter les changements et leurs effets
|
||||
4. Tester sur différents types de requêtes
|
||||
5. Créer des modèles spécialisés pour différents cas d'usage
|
||||
|
||||
## 7. Optimisation et intégration dans VSCode/Cursor
|
||||
|
||||
### 7.1 Installation et configuration de base
|
||||
|
||||
Pour intégrer Ollama dans VSCode/Cursor :
|
||||
|
||||
1. **Installer l'extension Continue** :
|
||||
```bash
|
||||
code --install-extension Continue.continue
|
||||
```
|
||||
|
||||
2. **Configurer l'extension** :
|
||||
Ouvrir les paramètres VSCode et configurer Continue :
|
||||
```json
|
||||
{
|
||||
"continue.model": "ollama",
|
||||
"continue.modelParameters": {
|
||||
"model": "codellama"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### 7.2 Utilisation dans VSCode/Cursor
|
||||
|
||||
#### Commandes principales :
|
||||
- Utiliser `Ctrl+Shift+P` (ou `Cmd+Shift+P` sur macOS) puis "Continue: Edit Code" pour générer du code
|
||||
- Sélectionner du code et utiliser "Continue: Explain Code" pour obtenir des explications
|
||||
- Utiliser "Continue: Edit Code" sur du code existant pour le refactoriser
|
||||
|
||||
#### Modèles recommandés pour VSCode/Cursor :
|
||||
- **Autocomplétion** : `starcoder2:3b`
|
||||
- **Génération de code** : `codellama` ou `deepseek-coder:6.7b`
|
||||
- **Débug et explication** : `llama3.1`
|
||||
|
||||
### 7.3 Configuration avancée pour Cursor
|
||||
|
||||
Pour Cursor, qui est basé sur VSCode, vous pouvez optimiser davantage :
|
||||
|
||||
1. **Configuration pour les projets de grande taille** :
|
||||
```json
|
||||
{
|
||||
"continue.model": "ollama",
|
||||
"continue.modelParameters": {
|
||||
"model": "codellama",
|
||||
"num_ctx": 8192,
|
||||
"temperature": 0.2
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
2. **Pour l'utilisation en ligne de commande** :
|
||||
```bash
|
||||
# Lancer le serveur Ollama
|
||||
ollama serve
|
||||
|
||||
# Exécuter le modèle
|
||||
ollama run codellama
|
||||
```
|
||||
|
||||
3. **Commandes utiles dans l'interface Continue** :
|
||||
- `"""` pour les messages multiligne
|
||||
- `/bye` pour quitter le modèle
|
||||
|
||||
### 7.4 Résolution des problèmes courants
|
||||
|
||||
Si vous rencontrez des problèmes :
|
||||
|
||||
1. **Vérifier que le serveur Ollama est en cours d'exécution** :
|
||||
```bash
|
||||
ps aux | grep ollama
|
||||
```
|
||||
|
||||
2. **Redémarrer le serveur** :
|
||||
```bash
|
||||
killall ollama
|
||||
ollama serve
|
||||
```
|
||||
|
||||
3. **Problèmes de performance** :
|
||||
- Utiliser un modèle plus petit
|
||||
- Fermer les applications gourmandes en ressources
|
||||
- Vérifier la disponibilité de la RAM
|
||||
|
||||
4. **Si l'extension ne répond pas** :
|
||||
- Recharger la fenêtre VSCode/Cursor
|
||||
- Vérifier l'état du serveur Ollama
|
||||
- Vérifier les paramètres de configuration
|
||||
BIN
images/143414_image.png
Normal file
BIN
images/143414_image.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 36 KiB |
8
llm_classes.egg-info/PKG-INFO
Normal file
8
llm_classes.egg-info/PKG-INFO
Normal file
@ -0,0 +1,8 @@
|
||||
Metadata-Version: 2.4
|
||||
Name: llm_classes
|
||||
Version: 0.1
|
||||
Requires-Dist: requests>=2.28.0
|
||||
Requires-Dist: datetime>=4.3
|
||||
Requires-Dist: pillow>=9.0.0
|
||||
Requires-Dist: argparse>=1.4.0
|
||||
Dynamic: requires-dist
|
||||
18
llm_classes.egg-info/SOURCES.txt
Normal file
18
llm_classes.egg-info/SOURCES.txt
Normal file
@ -0,0 +1,18 @@
|
||||
setup.py
|
||||
llm_classes/__init__.py
|
||||
llm_classes/agents.py
|
||||
llm_classes/deepl.py
|
||||
llm_classes/deepseek.py
|
||||
llm_classes/example.py
|
||||
llm_classes/example_vision.py
|
||||
llm_classes/llama_vision.py
|
||||
llm_classes/llm.py
|
||||
llm_classes/mistral.py
|
||||
llm_classes/ollama.py
|
||||
llm_classes/perplexity.py
|
||||
llm_classes/rag.py
|
||||
llm_classes.egg-info/PKG-INFO
|
||||
llm_classes.egg-info/SOURCES.txt
|
||||
llm_classes.egg-info/dependency_links.txt
|
||||
llm_classes.egg-info/requires.txt
|
||||
llm_classes.egg-info/top_level.txt
|
||||
1
llm_classes.egg-info/dependency_links.txt
Normal file
1
llm_classes.egg-info/dependency_links.txt
Normal file
@ -0,0 +1 @@
|
||||
|
||||
4
llm_classes.egg-info/requires.txt
Normal file
4
llm_classes.egg-info/requires.txt
Normal file
@ -0,0 +1,4 @@
|
||||
requests>=2.28.0
|
||||
datetime>=4.3
|
||||
pillow>=9.0.0
|
||||
argparse>=1.4.0
|
||||
1
llm_classes.egg-info/top_level.txt
Normal file
1
llm_classes.egg-info/top_level.txt
Normal file
@ -0,0 +1 @@
|
||||
llm_classes
|
||||
141
llm_classes/README.md
Normal file
141
llm_classes/README.md
Normal file
@ -0,0 +1,141 @@
|
||||
# Classes Python pour l'intégration de LLM
|
||||
|
||||
Ce package fournit des classes Python permettant d'intégrer facilement différents modèles de langage (LLM) dans vos applications.
|
||||
|
||||
## Classes disponibles
|
||||
|
||||
- `LLM` : Classe de base abstraite pour tous les modèles de langage
|
||||
- `Mistral` : Intégration avec l'API Mistral
|
||||
- `Ollama` : Intégration avec Ollama (modèles locaux)
|
||||
- `DeepSeek` : Intégration avec l'API DeepSeek
|
||||
- `Perplexity` : Intégration avec l'API Perplexity
|
||||
- `DeepL` : Intégration avec l'API DeepL pour la traduction
|
||||
- `RAG` : Système de Retrieval Augmented Generation
|
||||
- `LlamaVision` : Intégration avec Llama Vision 3.2 pour l'analyse d'images
|
||||
- `Agent`, `AgentAnalyseImage`, `AgentAnalyseJSON`, `AgentQuestionReponse` : Agents spécialisés pour l'analyse d'images et de données
|
||||
|
||||
## Installation
|
||||
|
||||
```bash
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
## Exemples d'utilisation
|
||||
|
||||
### Utilisation de Mistral
|
||||
|
||||
```python
|
||||
from llm_classes import Mistral
|
||||
|
||||
# Création d'une instance
|
||||
mistral = Mistral(api_key="votre_clé_api")
|
||||
|
||||
# Configuration
|
||||
mistral.temperature = 0.7
|
||||
mistral.max_tokens = 500
|
||||
|
||||
# Interrogation
|
||||
reponse = mistral.Interroger("Explique-moi la théorie de la relativité")
|
||||
print(reponse)
|
||||
```
|
||||
|
||||
### Utilisation d'Ollama
|
||||
|
||||
```python
|
||||
from llm_classes import Ollama
|
||||
|
||||
# Création d'une instance
|
||||
ollama = Ollama()
|
||||
ollama.Modele = "llama2" # ou tout autre modèle disponible dans Ollama
|
||||
|
||||
# Interrogation
|
||||
reponse = ollama.Interroger("Comment fonctionne une machine à vapeur?")
|
||||
print(reponse)
|
||||
```
|
||||
|
||||
### Utilisation de DeepL
|
||||
|
||||
```python
|
||||
from llm_classes import DeepL
|
||||
|
||||
# Création d'une instance
|
||||
deepl = DeepL(api_key="votre_clé_api")
|
||||
|
||||
# Traduction
|
||||
texte_traduit = deepl.traduire("Hello world", "FR")
|
||||
print(texte_traduit) # Bonjour le monde
|
||||
```
|
||||
|
||||
### Utilisation de LlamaVision pour l'analyse d'images
|
||||
|
||||
```python
|
||||
from llm_classes import LlamaVision
|
||||
|
||||
# Création d'une instance
|
||||
vision = LlamaVision()
|
||||
|
||||
# Chargement d'une image
|
||||
vision.set_image("chemin/vers/image.jpg")
|
||||
|
||||
# Analyse de l'image
|
||||
reponse = vision.Interroger("Décris en détail ce que tu vois dans cette image")
|
||||
print(reponse)
|
||||
|
||||
# Avec données JSON complémentaires
|
||||
vision.set_json_data({"contexte": "Photo prise en vacances"})
|
||||
reponse = vision.Interroger("En tenant compte du contexte, décris cette image")
|
||||
print(reponse)
|
||||
|
||||
# Sauvegarde des résultats
|
||||
vision.sauvegarder_resultats("resultats.json")
|
||||
```
|
||||
|
||||
### Utilisation des agents pour un workflow complet
|
||||
|
||||
```python
|
||||
from llm_classes import AgentAnalyseImage, AgentAnalyseJSON, AgentQuestionReponse
|
||||
|
||||
# Analyse d'une image
|
||||
agent_image = AgentAnalyseImage("Agent Vision")
|
||||
resultats_image = agent_image.executer("chemin/vers/image.jpg")
|
||||
|
||||
# Analyse des données extraites
|
||||
agent_json = AgentAnalyseJSON("Agent JSON")
|
||||
resultats_analyse = agent_json.executer(resultats_image, "Analyse les éléments clés de cette image")
|
||||
|
||||
# Questions/Réponses sur les données
|
||||
agent_qr = AgentQuestionReponse("Agent QR")
|
||||
questions = [
|
||||
"Que représente principalement cette image?",
|
||||
"Y a-t-il des personnes ou des animaux visibles?"
|
||||
]
|
||||
resultats_qr = agent_qr.executer(resultats_analyse, questions)
|
||||
|
||||
# Évaluation des réponses
|
||||
evaluation = agent_qr.evaluer_reponses(resultats_qr["resultats_qr"])
|
||||
print(f"Note moyenne des réponses: {evaluation['meta']['note_moyenne']}/10")
|
||||
```
|
||||
|
||||
## Scripts d'exemple
|
||||
|
||||
Le package inclut deux scripts d'exemple :
|
||||
|
||||
- `example.py` : Démontre l'utilisation des classes Mistral, Ollama, DeepSeek, Perplexity, DeepL et RAG
|
||||
- `example_vision.py` : Démontre l'utilisation de LlamaVision et des agents d'analyse d'images
|
||||
|
||||
### Utilisation de example_vision.py
|
||||
|
||||
```bash
|
||||
# Analyse directe d'une image
|
||||
python example_vision.py --image chemin/vers/image.jpg --mode direct
|
||||
|
||||
# Workflow complet avec analyse et génération de questions/réponses
|
||||
python example_vision.py --image chemin/vers/image.jpg --json donnees_complementaires.json
|
||||
```
|
||||
|
||||
## Notes importantes
|
||||
|
||||
1. Pour utiliser ces classes, vous devez disposer des clés API appropriées pour les services externes (Mistral, DeepSeek, Perplexity, DeepL).
|
||||
2. Pour LlamaVision et Ollama, assurez-vous que les services correspondants sont accessibles (généralement sur localhost).
|
||||
3. Certaines fonctionnalités nécessitent des dépendances supplémentaires comme Pillow pour le traitement d'images.
|
||||
4. Les exemples supposent que les services sont correctement configurés et accessibles.
|
||||
15
llm_classes/__init__.py
Normal file
15
llm_classes/__init__.py
Normal file
@ -0,0 +1,15 @@
|
||||
# Importation des classes pour faciliter l'accès
|
||||
from .llm import LLM
|
||||
from .mistral import Mistral
|
||||
from .ollama import Ollama
|
||||
from .deepseek import DeepSeek
|
||||
from .perplexity import Perplexity
|
||||
from .deepl import DeepL
|
||||
from .rag import RAG
|
||||
from .llama_vision import LlamaVision
|
||||
from .agents import Agent, AgentAnalyseImage, AgentAnalyseJSON, AgentQuestionReponse
|
||||
|
||||
__all__ = [
|
||||
'LLM', 'Mistral', 'Ollama', 'DeepSeek', 'Perplexity', 'DeepL', 'RAG',
|
||||
'LlamaVision', 'Agent', 'AgentAnalyseImage', 'AgentAnalyseJSON', 'AgentQuestionReponse'
|
||||
]
|
||||
BIN
llm_classes/__pycache__/__init__.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/__init__.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/agents.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/agents.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/deepl.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/deepl.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/deepseek.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/deepseek.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/llama_vision.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/llama_vision.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/llm.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/llm.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/mistral.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/mistral.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/ollama.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/ollama.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/perplexity.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/perplexity.cpython-312.pyc
Normal file
Binary file not shown.
BIN
llm_classes/__pycache__/rag.cpython-312.pyc
Normal file
BIN
llm_classes/__pycache__/rag.cpython-312.pyc
Normal file
Binary file not shown.
436
llm_classes/agents.py
Normal file
436
llm_classes/agents.py
Normal file
@ -0,0 +1,436 @@
|
||||
import json
|
||||
import os
|
||||
from typing import Dict, List, Any, Optional, Union, Tuple
|
||||
import time
|
||||
from datetime import datetime
|
||||
from .llama_vision import LlamaVision
|
||||
from .mistral import Mistral
|
||||
from .ollama import Ollama
|
||||
|
||||
class Agent:
|
||||
"""
|
||||
Classe de base pour tous les agents
|
||||
"""
|
||||
def __init__(self, nom: str = "Agent"):
|
||||
self.nom: str = nom
|
||||
self.historique: List[Dict[str, Any]] = []
|
||||
|
||||
def ajouter_historique(self, action: str, input_data: Any, output_data: Any) -> None:
|
||||
"""
|
||||
Ajoute une entrée dans l'historique de l'agent
|
||||
"""
|
||||
self.historique.append({
|
||||
"timestamp": datetime.now().isoformat(),
|
||||
"action": action,
|
||||
"input": str(input_data)[:500], # Limite pour éviter des historiques trop grands
|
||||
"output": str(output_data)[:500] # Limite pour éviter des historiques trop grands
|
||||
})
|
||||
|
||||
def obtenir_historique(self) -> List[Dict[str, Any]]:
|
||||
"""
|
||||
Retourne l'historique complet de l'agent
|
||||
"""
|
||||
return self.historique
|
||||
|
||||
def executer(self, *args, **kwargs) -> Any:
|
||||
"""
|
||||
Méthode abstraite à implémenter dans les classes dérivées
|
||||
"""
|
||||
raise NotImplementedError("Chaque agent doit implémenter sa propre méthode executer()")
|
||||
|
||||
|
||||
class AgentAnalyseImage(Agent):
|
||||
"""
|
||||
Agent pour analyser des images avec Llama Vision
|
||||
"""
|
||||
def __init__(self, nom: str = "AgentAnalyseImage"):
|
||||
super().__init__(nom)
|
||||
self.llm = LlamaVision()
|
||||
self.questions_standard = [
|
||||
"Décris en détail ce que tu vois sur cette image.",
|
||||
"Quels sont les éléments principaux visibles sur cette image?",
|
||||
"Y a-t-il du texte visible sur cette image? Si oui, peux-tu le transcrire?",
|
||||
"Quelle est l'ambiance générale de cette image?"
|
||||
]
|
||||
|
||||
def executer(self, image_path: str, json_data: Optional[Dict[str, Any]] = None,
|
||||
question: Optional[str] = None) -> Dict[str, Any]:
|
||||
"""
|
||||
Analyse une image avec Llama Vision
|
||||
|
||||
Args:
|
||||
image_path: Chemin vers l'image à analyser
|
||||
json_data: Données JSON optionnelles associées à l'image
|
||||
question: Question à poser au modèle (si None, utilise les questions standard)
|
||||
|
||||
Returns:
|
||||
Résultats de l'analyse sous forme de dictionnaire
|
||||
"""
|
||||
# Vérification de l'existence de l'image
|
||||
if not os.path.exists(image_path):
|
||||
resultat = {"erreur": f"L'image {image_path} n'existe pas"}
|
||||
self.ajouter_historique("analyse_image_erreur", image_path, resultat)
|
||||
return resultat
|
||||
|
||||
# Chargement de l'image
|
||||
if not self.llm.set_image(image_path):
|
||||
resultat = {"erreur": f"Échec du chargement de l'image {image_path}"}
|
||||
self.ajouter_historique("analyse_image_erreur", image_path, resultat)
|
||||
return resultat
|
||||
|
||||
# Ajout des données JSON si fournies
|
||||
if json_data:
|
||||
self.llm.set_json_data(json_data)
|
||||
|
||||
# Exécution de l'analyse
|
||||
resultats = {}
|
||||
|
||||
# Utilisation des questions standard si aucune question n'est fournie
|
||||
questions_a_poser = [question] if question else self.questions_standard
|
||||
|
||||
for idx, q in enumerate(questions_a_poser):
|
||||
print(f"Question {idx+1}/{len(questions_a_poser)}: {q[:50]}...")
|
||||
reponse = self.llm.Interroger(q)
|
||||
resultats[f"question_{idx+1}"] = {
|
||||
"question": q,
|
||||
"reponse": reponse
|
||||
}
|
||||
time.sleep(1) # Pause pour éviter de surcharger l'API
|
||||
|
||||
# Fusion avec le JSON si présent
|
||||
resultats_complets = self.llm.fusionner_json_avec_resultats()
|
||||
resultats_complets["analyses"] = resultats
|
||||
|
||||
# Ajout à l'historique
|
||||
self.ajouter_historique("analyse_image", image_path, "Analyse complétée avec succès")
|
||||
|
||||
return resultats_complets
|
||||
|
||||
def ajuster_parametres(self, temperature: Optional[float] = None,
|
||||
top_p: Optional[float] = None,
|
||||
num_ctx: Optional[int] = None) -> None:
|
||||
"""
|
||||
Ajuste les paramètres du modèle Llama Vision
|
||||
"""
|
||||
if temperature is not None:
|
||||
self.llm.o_temperature = temperature
|
||||
|
||||
if top_p is not None:
|
||||
self.llm.o_top_p = top_p
|
||||
|
||||
if num_ctx is not None:
|
||||
self.llm.o_num_ctx = num_ctx
|
||||
|
||||
self.ajouter_historique("ajuster_parametres",
|
||||
{"temperature": temperature, "top_p": top_p, "num_ctx": num_ctx},
|
||||
"Paramètres ajustés")
|
||||
|
||||
def sauvegarder_resultats(self, chemin_fichier: str, resultats: Dict[str, Any]) -> bool:
|
||||
"""
|
||||
Sauvegarde les résultats d'analyse dans un fichier JSON
|
||||
"""
|
||||
try:
|
||||
with open(chemin_fichier, 'w', encoding='utf-8') as f:
|
||||
json.dump(resultats, f, ensure_ascii=False, indent=2)
|
||||
|
||||
self.ajouter_historique("sauvegarder_resultats", chemin_fichier, "Résultats sauvegardés")
|
||||
return True
|
||||
except Exception as e:
|
||||
self.ajouter_historique("sauvegarder_resultats_erreur", chemin_fichier, str(e))
|
||||
return False
|
||||
|
||||
|
||||
class AgentAnalyseJSON(Agent):
|
||||
"""
|
||||
Agent pour analyser des données JSON
|
||||
"""
|
||||
def __init__(self, nom: str = "AgentAnalyseJSON", modele: str = "mistral"):
|
||||
super().__init__(nom)
|
||||
|
||||
# Choix du modèle
|
||||
if modele.lower() == "mistral":
|
||||
self.llm = Mistral()
|
||||
self.llm.prompt_system = "Tu es un expert en analyse de données JSON. Tu dois extraire des informations pertinentes, identifier des tendances et répondre à des questions sur les données."
|
||||
else:
|
||||
self.llm = Ollama()
|
||||
self.llm.prompt_system = "Tu es un expert en analyse de données JSON. Tu dois extraire des informations pertinentes, identifier des tendances et répondre à des questions sur les données."
|
||||
self.llm.Modele = modele
|
||||
|
||||
def executer(self, json_data: Dict[str, Any],
|
||||
question: str = "Analyse ces données et extrait les informations principales.") -> Dict[str, Any]:
|
||||
"""
|
||||
Analyse des données JSON
|
||||
|
||||
Args:
|
||||
json_data: Données JSON à analyser
|
||||
question: Question à poser au modèle
|
||||
|
||||
Returns:
|
||||
Résultats de l'analyse sous forme de dictionnaire
|
||||
"""
|
||||
# Conversion du JSON en chaîne formatée
|
||||
json_str = json.dumps(json_data, ensure_ascii=False, indent=2)
|
||||
|
||||
# Construction du prompt avec le JSON et la question
|
||||
prompt = f"{question}\n\nDonnées JSON à analyser:\n```json\n{json_str}\n```"
|
||||
|
||||
# Interrogation du modèle
|
||||
reponse = self.llm.Interroger(prompt)
|
||||
|
||||
# Construction du résultat
|
||||
resultats = {
|
||||
"question": question,
|
||||
"reponse": reponse,
|
||||
"timestamp": datetime.now().isoformat(),
|
||||
"taille_json": len(json_str)
|
||||
}
|
||||
|
||||
# Ajout à l'historique
|
||||
self.ajouter_historique("analyse_json", prompt[:200] + "...", reponse[:200] + "...")
|
||||
|
||||
return resultats
|
||||
|
||||
def extraire_structure(self, json_data: Dict[str, Any]) -> Dict[str, Any]:
|
||||
"""
|
||||
Extrait la structure d'un JSON (clés, types, profondeur)
|
||||
"""
|
||||
resultat = {
|
||||
"structure": {},
|
||||
"statistiques": {
|
||||
"nb_cles": 0,
|
||||
"profondeur_max": 0,
|
||||
"types": {}
|
||||
}
|
||||
}
|
||||
|
||||
def explorer_structure(data, chemin="", profondeur=0):
|
||||
nonlocal resultat
|
||||
|
||||
# Mise à jour de la profondeur max
|
||||
resultat["statistiques"]["profondeur_max"] = max(resultat["statistiques"]["profondeur_max"], profondeur)
|
||||
|
||||
if isinstance(data, dict):
|
||||
structure = {}
|
||||
for cle, valeur in data.items():
|
||||
nouveau_chemin = f"{chemin}.{cle}" if chemin else cle
|
||||
resultat["statistiques"]["nb_cles"] += 1
|
||||
|
||||
type_valeur = type(valeur).__name__
|
||||
if type_valeur not in resultat["statistiques"]["types"]:
|
||||
resultat["statistiques"]["types"][type_valeur] = 0
|
||||
resultat["statistiques"]["types"][type_valeur] += 1
|
||||
|
||||
if isinstance(valeur, (dict, list)):
|
||||
structure[cle] = explorer_structure(valeur, nouveau_chemin, profondeur + 1)
|
||||
else:
|
||||
structure[cle] = type_valeur
|
||||
return structure
|
||||
|
||||
elif isinstance(data, list):
|
||||
if data and isinstance(data[0], (dict, list)):
|
||||
# Pour les listes de structures complexes, on analyse le premier élément
|
||||
return [explorer_structure(data[0], f"{chemin}[0]", profondeur + 1)]
|
||||
else:
|
||||
# Pour les listes de valeurs simples
|
||||
type_elements = "vide" if not data else type(data[0]).__name__
|
||||
resultat["statistiques"]["nb_cles"] += 1
|
||||
|
||||
if "list" not in resultat["statistiques"]["types"]:
|
||||
resultat["statistiques"]["types"]["list"] = 0
|
||||
resultat["statistiques"]["types"]["list"] += 1
|
||||
|
||||
return f"list[{type_elements}]"
|
||||
else:
|
||||
return type(data).__name__
|
||||
|
||||
resultat["structure"] = explorer_structure(json_data)
|
||||
|
||||
self.ajouter_historique("extraire_structure", "JSON", resultat)
|
||||
return resultat
|
||||
|
||||
def fusionner_jsons(self, json1: Dict[str, Any], json2: Dict[str, Any]) -> Dict[str, Any]:
|
||||
"""
|
||||
Fusionne deux structures JSON en conservant les informations des deux
|
||||
"""
|
||||
if not json1:
|
||||
return json2
|
||||
|
||||
if not json2:
|
||||
return json1
|
||||
|
||||
resultat = json1.copy()
|
||||
|
||||
# Fonction récursive pour fusionner
|
||||
def fusionner(dict1, dict2):
|
||||
for cle, valeur in dict2.items():
|
||||
if cle in dict1:
|
||||
# Si les deux sont des dictionnaires, fusion récursive
|
||||
if isinstance(dict1[cle], dict) and isinstance(valeur, dict):
|
||||
fusionner(dict1[cle], valeur)
|
||||
# Si les deux sont des listes, concaténation
|
||||
elif isinstance(dict1[cle], list) and isinstance(valeur, list):
|
||||
dict1[cle].extend(valeur)
|
||||
# Sinon, on garde les deux valeurs dans une liste
|
||||
else:
|
||||
if not isinstance(dict1[cle], list):
|
||||
dict1[cle] = [dict1[cle]]
|
||||
if isinstance(valeur, list):
|
||||
dict1[cle].extend(valeur)
|
||||
else:
|
||||
dict1[cle].append(valeur)
|
||||
else:
|
||||
# Si la clé n'existe pas dans dict1, on l'ajoute simplement
|
||||
dict1[cle] = valeur
|
||||
|
||||
fusionner(resultat, json2)
|
||||
self.ajouter_historique("fusionner_jsons", "Fusion de deux JSON", "Fusion réussie")
|
||||
|
||||
return resultat
|
||||
|
||||
|
||||
class AgentQuestionReponse(Agent):
|
||||
"""
|
||||
Agent pour tester des questions/réponses sur des données JSON
|
||||
"""
|
||||
def __init__(self, nom: str = "AgentQuestionReponse", modele: str = "mistral"):
|
||||
super().__init__(nom)
|
||||
|
||||
# Choix du modèle
|
||||
if modele.lower() == "mistral":
|
||||
self.llm = Mistral()
|
||||
else:
|
||||
self.llm = Ollama()
|
||||
self.llm.Modele = modele
|
||||
|
||||
# Configuration du modèle
|
||||
self.llm.prompt_system = "Tu es un assistant expert qui répond à des questions en te basant uniquement sur les données JSON fournies. Ne fais pas de suppositions au-delà des données."
|
||||
|
||||
# Questions prédéfinies
|
||||
self.questions_standard = [
|
||||
"Quel est le sujet principal de cette image d'après les données?",
|
||||
"Quels sont les éléments clés identifiés dans l'image?",
|
||||
"Y a-t-il des incohérences entre les différentes analyses?",
|
||||
"Résume les informations principales contenues dans ce JSON."
|
||||
]
|
||||
|
||||
def executer(self, json_data: Dict[str, Any],
|
||||
questions: Optional[List[str]] = None) -> Dict[str, List[Dict[str, str]]]:
|
||||
"""
|
||||
Exécute une série de questions/réponses sur des données JSON
|
||||
|
||||
Args:
|
||||
json_data: Données JSON à analyser
|
||||
questions: Liste de questions à poser (si None, utilise les questions standard)
|
||||
|
||||
Returns:
|
||||
Résultats des Q/R sous forme de dictionnaire
|
||||
"""
|
||||
questions_a_poser = questions if questions else self.questions_standard
|
||||
resultats = []
|
||||
|
||||
# Conversion du JSON en chaîne formatée
|
||||
json_str = json.dumps(json_data, ensure_ascii=False, indent=2)
|
||||
|
||||
for question in questions_a_poser:
|
||||
# Construction du prompt
|
||||
prompt = f"Question: {question}\n\nVoici les données JSON sur lesquelles te baser pour répondre:\n```json\n{json_str}\n```\n\nRéponds de manière claire et concise en te basant uniquement sur ces données."
|
||||
|
||||
# Interrogation du modèle
|
||||
reponse = self.llm.Interroger(prompt)
|
||||
|
||||
resultats.append({
|
||||
"question": question,
|
||||
"reponse": reponse
|
||||
})
|
||||
|
||||
# Ajout à l'historique
|
||||
self.ajouter_historique("question_reponse", question, reponse[:200] + "...")
|
||||
|
||||
# Pause pour éviter de surcharger l'API
|
||||
time.sleep(1)
|
||||
|
||||
return {"resultats_qr": resultats}
|
||||
|
||||
def generer_questions(self, json_data: Dict[str, Any], nb_questions: int = 3) -> List[str]:
|
||||
"""
|
||||
Génère automatiquement des questions pertinentes basées sur les données JSON
|
||||
|
||||
Args:
|
||||
json_data: Données JSON à analyser
|
||||
nb_questions: Nombre de questions à générer
|
||||
|
||||
Returns:
|
||||
Liste de questions générées
|
||||
"""
|
||||
# Conversion du JSON en chaîne formatée
|
||||
json_str = json.dumps(json_data, ensure_ascii=False, indent=2)
|
||||
|
||||
prompt = f"Voici des données JSON contenant des analyses d'images:\n```json\n{json_str}\n```\n\nGénère {nb_questions} questions pertinentes et spécifiques que l'on pourrait poser à propos de ces données. Donne uniquement la liste des questions, une par ligne."
|
||||
|
||||
reponse = self.llm.Interroger(prompt)
|
||||
|
||||
# Extraction des questions (chaque ligne est une question)
|
||||
questions = [ligne.strip() for ligne in reponse.split('\n') if ligne.strip()]
|
||||
|
||||
# Filtrage des lignes qui ne sont pas des questions
|
||||
questions = [q for q in questions if '?' in q][:nb_questions]
|
||||
|
||||
self.ajouter_historique("generer_questions", f"Génération de {nb_questions} questions", questions)
|
||||
|
||||
return questions
|
||||
|
||||
def evaluer_reponses(self, resultats_qr: List[Dict[str, str]]) -> Dict[str, Any]:
|
||||
"""
|
||||
Évalue la qualité des réponses générées
|
||||
|
||||
Args:
|
||||
resultats_qr: Liste de dictionnaires avec des paires question/réponse
|
||||
|
||||
Returns:
|
||||
Évaluation des réponses
|
||||
"""
|
||||
evaluation = {
|
||||
"meta": {
|
||||
"nb_questions": len(resultats_qr),
|
||||
"timestamp": datetime.now().isoformat()
|
||||
},
|
||||
"evaluations": []
|
||||
}
|
||||
|
||||
for idx, qr in enumerate(resultats_qr):
|
||||
question = qr["question"]
|
||||
reponse = qr["reponse"]
|
||||
|
||||
# Construction du prompt d'évaluation
|
||||
prompt = f"Évalue la qualité de cette réponse à la question suivante. Attribue une note de 1 à 10 et explique pourquoi.\n\nQuestion: {question}\n\nRéponse: {reponse}\n\nFormat de réponse attendu: Note: [1-10]\nJustification: [explication]"
|
||||
|
||||
eval_reponse = self.llm.Interroger(prompt)
|
||||
|
||||
# Extraction de la note
|
||||
note = 0
|
||||
for ligne in eval_reponse.split('\n'):
|
||||
if ligne.lower().startswith('note:'):
|
||||
try:
|
||||
# Extraction de la partie numérique
|
||||
note_str = ''.join(c for c in ligne if c.isdigit() or c == '.')
|
||||
note = float(note_str)
|
||||
break
|
||||
except:
|
||||
note = 0
|
||||
|
||||
evaluation["evaluations"].append({
|
||||
"index": idx,
|
||||
"question": question,
|
||||
"note": note,
|
||||
"evaluation": eval_reponse
|
||||
})
|
||||
|
||||
time.sleep(1) # Pause pour éviter de surcharger l'API
|
||||
|
||||
# Calcul de la note moyenne
|
||||
notes = [e["note"] for e in evaluation["evaluations"]]
|
||||
evaluation["meta"]["note_moyenne"] = sum(notes) / len(notes) if notes else 0
|
||||
|
||||
self.ajouter_historique("evaluer_reponses", "Évaluation des réponses", evaluation["meta"])
|
||||
|
||||
return evaluation
|
||||
272
llm_classes/deepl.py
Normal file
272
llm_classes/deepl.py
Normal file
@ -0,0 +1,272 @@
|
||||
import json
|
||||
from typing import Dict, List, Any, Optional, Union, Tuple
|
||||
from .llm import LLM
|
||||
|
||||
class DeepL(LLM):
|
||||
"""
|
||||
Classe pour l'intégration avec l'API DeepL
|
||||
Cette classe hérite de la classe de base LLM et permet de faire des traductions
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""
|
||||
Initialisation des attributs spécifiques à DeepL
|
||||
"""
|
||||
super().__init__()
|
||||
|
||||
# Attributs spécifiques à DeepL
|
||||
self.langueSource: str = "FR"
|
||||
self.langueDesti: str = "EN"
|
||||
self.context: str = ""
|
||||
self.split_sentences: str = "0"
|
||||
self.preserve_formatting: bool = True
|
||||
self.formality: str = "default" # Options: more, less, prefer_more, prefer_less
|
||||
self.show_billed_characters: bool = False
|
||||
self.tag_handling: str = "" # Options: xml, html
|
||||
self.outline_detection: bool = True
|
||||
self.non_splitting_tags: List[str] = []
|
||||
self.splitting_tags: List[str] = []
|
||||
self.ignore_tags: List[str] = []
|
||||
|
||||
# Constantes publiques (modifiées en attributs en Python)
|
||||
self.glossaire_id: Dict[str, str] = {} # tableau associatif (ccSansCasse) de chaînes
|
||||
self._m_tabGlossaireListePaire: List[List[str]] = [] # tableau de 0 par 2 chaînes
|
||||
|
||||
# Initialisation par défaut
|
||||
self.langueSource = "FR"
|
||||
self.langueDesti = "EN"
|
||||
self.Modele = "prefer_quality_optimized"
|
||||
self.split_sentences = "0"
|
||||
self.preserve_formatting = True
|
||||
self.formality = "default"
|
||||
self.show_billed_characters = False
|
||||
self.tag_handling = ""
|
||||
self.outline_detection = True
|
||||
|
||||
def commandeAutorisation(self) -> str:
|
||||
"""
|
||||
Retourne la commande d'autorisation pour DeepL
|
||||
"""
|
||||
return "DeepL-Auth-Key"
|
||||
|
||||
def cleAPI(self) -> str:
|
||||
"""
|
||||
Retourne la clé API DeepL
|
||||
"""
|
||||
return "d42ee2f7-f4e6-d437-3237-fa40b4f45e65"
|
||||
|
||||
def urlBase(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de base de l'API DeepL
|
||||
"""
|
||||
return "https://api.deepl.com/v2/"
|
||||
|
||||
def urlFonction(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de la fonction par défaut (translate)
|
||||
"""
|
||||
return "translate"
|
||||
|
||||
def model_list(self) -> List[str]:
|
||||
"""
|
||||
Retourne la liste des modèles disponibles pour DeepL
|
||||
"""
|
||||
return ["quality_optimized", "latency_optimized", "prefer_quality_optimized"]
|
||||
|
||||
def glossaire_liste_paire_possible(self) -> List[List[str]]:
|
||||
"""
|
||||
Récupère la liste des paires de langues possibles pour les glossaires
|
||||
"""
|
||||
import requests
|
||||
|
||||
if len(self._m_tabGlossaireListePaire) > 1:
|
||||
return self._m_tabGlossaireListePaire
|
||||
|
||||
url = self.urlBase() + "glossary-language-pairs"
|
||||
headers = {
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
if self.cleAPI() != "":
|
||||
headers["Authorization"] = f"{self.commandeAutorisation()} {self.cleAPI()}"
|
||||
|
||||
try:
|
||||
response = requests.get(url=url, headers=headers)
|
||||
|
||||
if response.status_code == 200:
|
||||
data = json.loads(response.text)
|
||||
self._m_tabGlossaireListePaire = []
|
||||
|
||||
for item in data["supported_languages"]:
|
||||
self._m_tabGlossaireListePaire.append([
|
||||
item["source_lang"],
|
||||
item["target_lang"]
|
||||
])
|
||||
else:
|
||||
self._m_tabGlossaireListePaire.append([response.text])
|
||||
except Exception as e:
|
||||
self._m_tabGlossaireListePaire.append([str(e)])
|
||||
|
||||
return self._m_tabGlossaireListePaire
|
||||
|
||||
def _even_LLM_POST(self, question: str) -> None:
|
||||
"""
|
||||
Préparation du contenu de la requête pour DeepL
|
||||
"""
|
||||
# Paramètre principal - le texte à traduire
|
||||
self._Contenu["text"] = [question]
|
||||
self._Contenu["source_lang"] = self.langueSource
|
||||
self._Contenu["target_lang"] = self.langueDesti
|
||||
|
||||
# Ajout du contexte si défini
|
||||
if self.context:
|
||||
self._Contenu["context"] = self.context
|
||||
|
||||
# Définition du modèle
|
||||
self._Contenu["model_type"] = self.Modele
|
||||
|
||||
# Options supplémentaires
|
||||
self._Contenu["split_sentences"] = self.split_sentences
|
||||
self._Contenu["preserve_formatting"] = self.preserve_formatting
|
||||
self._Contenu["formality"] = self.formality
|
||||
|
||||
# Ajout du glossaire si disponible
|
||||
glossaire_id = self._glossaireRetrouveID(self.langueSource, self.langueDesti)
|
||||
if glossaire_id:
|
||||
self._Contenu["glossary_id"] = glossaire_id
|
||||
|
||||
# Options avancées
|
||||
self._Contenu["show_billed_characters"] = self.show_billed_characters
|
||||
self._Contenu["tag_handling"] = self.tag_handling
|
||||
self._Contenu["outline_detection"] = self.outline_detection
|
||||
|
||||
# Ajout des tags non-splittants si définis
|
||||
if self.non_splitting_tags:
|
||||
self._Contenu["non_splitting_tags"] = self.non_splitting_tags
|
||||
|
||||
# Ajout des tags splittants si définis
|
||||
if self.splitting_tags:
|
||||
self._Contenu["splitting_tags"] = self.splitting_tags
|
||||
|
||||
# Ajout des tags à ignorer si définis
|
||||
if self.ignore_tags:
|
||||
self._Contenu["ignore_tags"] = self.ignore_tags
|
||||
|
||||
def _interrogerRetourneReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Extraction de la réponse à partir du JSON retourné par DeepL
|
||||
"""
|
||||
data = json.loads(reponse)
|
||||
return data["translations"][0]["text"]
|
||||
|
||||
def glossaire_cree(self, sLanguageSource: str, sLangageDesti: str, sEntrees: str, sFormat: str = "tsv") -> str:
|
||||
"""
|
||||
Crée un glossaire avec les entrées fournies
|
||||
Format des entrées:
|
||||
- tsv : source<tab>cible<RC>
|
||||
- csv : source,cible<RC>
|
||||
"""
|
||||
import requests
|
||||
|
||||
# Supprime le glossaire existant si présent
|
||||
sID = self._glossaireRetrouveID(sLanguageSource, sLangageDesti)
|
||||
if sID:
|
||||
self._glossaireSupprime(sID)
|
||||
|
||||
url = self.urlBase() + "glossaries"
|
||||
headers = {
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
if self.cleAPI():
|
||||
headers["Authorization"] = f"{self.commandeAutorisation()} {self.cleAPI()}"
|
||||
|
||||
# Préparation des paramètres
|
||||
params = {
|
||||
"name": f"{sLanguageSource.upper()}_{sLangageDesti.upper()}",
|
||||
"entries": sEntrees,
|
||||
"source_lang": sLanguageSource,
|
||||
"target_lang": sLangageDesti,
|
||||
"entries_format": sFormat
|
||||
}
|
||||
|
||||
try:
|
||||
response = requests.post(url=url, headers=headers, json=params)
|
||||
|
||||
if response.status_code == 201:
|
||||
data = json.loads(response.text)
|
||||
return data["glossary_id"]
|
||||
else:
|
||||
return response.text
|
||||
except Exception as e:
|
||||
return str(e)
|
||||
|
||||
def _glossaireRetrouveID(self, sLanguageSource: str, sLangageDesti: str) -> str:
|
||||
"""
|
||||
Récupère l'ID d'un glossaire existant pour une paire de langues
|
||||
"""
|
||||
import requests
|
||||
|
||||
# Création de la désignation pour la paire de langues
|
||||
designation = f"{sLanguageSource.upper()}_{sLangageDesti.upper()}"
|
||||
|
||||
# Retourne l'ID du glossaire si déjà en cache
|
||||
if designation in self.glossaire_id:
|
||||
return self.glossaire_id[designation]
|
||||
|
||||
url = self.urlBase() + "glossaries"
|
||||
headers = {
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
if self.cleAPI():
|
||||
headers["Authorization"] = f"{self.commandeAutorisation()} {self.cleAPI()}"
|
||||
|
||||
try:
|
||||
response = requests.get(url=url, headers=headers)
|
||||
|
||||
if response.status_code == 200:
|
||||
data = json.loads(response.text)
|
||||
|
||||
# Mise à jour du cache des glossaires
|
||||
self.glossaire_id.clear()
|
||||
for glossaire in data["glossaries"]:
|
||||
glossaire_designation = f"{glossaire['source_lang'].upper()}_{glossaire['target_lang'].upper()}"
|
||||
self.glossaire_id[glossaire_designation] = glossaire["glossary_id"]
|
||||
|
||||
# Retourne l'ID du glossaire demandé
|
||||
designation = f"{sLanguageSource.upper()}_{sLangageDesti.upper()}"
|
||||
return self.glossaire_id.get(designation, "")
|
||||
else:
|
||||
return response.text
|
||||
except Exception as e:
|
||||
return str(e)
|
||||
|
||||
def _glossaireSupprime(self, sID: str) -> str:
|
||||
"""
|
||||
Supprime un glossaire existant
|
||||
"""
|
||||
import requests
|
||||
|
||||
url = self.urlBase() + f"glossaries/{sID}"
|
||||
headers = {
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
if self.cleAPI():
|
||||
headers["Authorization"] = f"{self.commandeAutorisation()} {self.cleAPI()}"
|
||||
|
||||
try:
|
||||
response = requests.delete(url=url, headers=headers)
|
||||
|
||||
# Suppression de l'entrée dans le cache
|
||||
for designation, id_en_cours in list(self.glossaire_id.items()):
|
||||
if id_en_cours == sID:
|
||||
del self.glossaire_id[designation]
|
||||
|
||||
if response.status_code == 204:
|
||||
return ""
|
||||
else:
|
||||
return response.text
|
||||
except Exception as e:
|
||||
return str(e)
|
||||
87
llm_classes/deepseek.py
Normal file
87
llm_classes/deepseek.py
Normal file
@ -0,0 +1,87 @@
|
||||
import json
|
||||
from typing import Dict, List, Any, Optional, Union
|
||||
from .llm import LLM
|
||||
|
||||
class DeepSeek(LLM):
|
||||
"""
|
||||
Classe pour l'intégration avec l'API DeepSeek
|
||||
Cette classe hérite de la classe de base LLM
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""
|
||||
Initialisation des attributs spécifiques à DeepSeek
|
||||
"""
|
||||
super().__init__()
|
||||
|
||||
# Attributs spécifiques à DeepSeek
|
||||
self.frequency_penalty: int = 0
|
||||
self.max_tokens: int = 4096
|
||||
self.presence_penalty: int = 0
|
||||
self.logprobs: bool = False
|
||||
self.top_logprobs: int = 0
|
||||
self.tool_choice: str = ""
|
||||
|
||||
# Initialisation par défaut
|
||||
self.Modele = "deepseek-chat"
|
||||
self.frequency_penalty = 0
|
||||
self.max_tokens = 4096
|
||||
self.presence_penalty = 0
|
||||
self.o_top_p = 1
|
||||
self.logprobs = False
|
||||
self.top_logprobs = 0
|
||||
self.tool_choice = ""
|
||||
|
||||
def urlBase(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de base de l'API DeepSeek
|
||||
"""
|
||||
return "https://api.deepseek.com/"
|
||||
|
||||
def cleAPI(self) -> str:
|
||||
"""
|
||||
Retourne la clé API DeepSeek
|
||||
"""
|
||||
return "sk-d359d9236ca84a5986f889631832d1e6"
|
||||
|
||||
def urlFonction(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de la fonction par défaut pour DeepSeek
|
||||
"""
|
||||
return "chat/completions"
|
||||
|
||||
def model_list(self) -> List[str]:
|
||||
"""
|
||||
Retourne la liste des modèles disponibles pour DeepSeek
|
||||
"""
|
||||
return ["deepseek-chat", "deepseek-reasoner"]
|
||||
|
||||
def _even_LLM_POST(self, question: str) -> None:
|
||||
"""
|
||||
Préparation du contenu de la requête pour DeepSeek
|
||||
"""
|
||||
self._Contenu["messages"] = [
|
||||
{"role": "system", "content": self.prompt_system},
|
||||
{"role": "user", "content": question}
|
||||
]
|
||||
|
||||
self._Contenu["stream"] = self._stream
|
||||
self._Contenu["frequency_penalty"] = self.frequency_penalty
|
||||
self._Contenu["max_tokens"] = self.max_tokens
|
||||
self._Contenu["presence_penalty"] = self.presence_penalty
|
||||
|
||||
# Ajout des logprobs si activés
|
||||
if self.logprobs:
|
||||
self._Contenu["logprobs"] = self.logprobs
|
||||
self._Contenu["top_logprobs"] = self.top_logprobs
|
||||
|
||||
# Ajout du tool_choice si défini
|
||||
if self.tool_choice:
|
||||
self._Contenu["tool_choice"] = self.tool_choice
|
||||
|
||||
def _interrogerRetourneReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Extraction de la réponse à partir du JSON retourné par DeepSeek
|
||||
"""
|
||||
data = json.loads(reponse)
|
||||
return data["choices"][0]["message"]["content"]
|
||||
137
llm_classes/example.py
Normal file
137
llm_classes/example.py
Normal file
@ -0,0 +1,137 @@
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
"""
|
||||
Script d'exemple pour tester les différentes classes LLM
|
||||
Ce script montre comment initialiser et utiliser chacune des classes
|
||||
"""
|
||||
|
||||
from llm_classes import Mistral, Ollama, DeepSeek, Perplexity, DeepL, RAG
|
||||
|
||||
def test_mistral():
|
||||
"""Test de la classe Mistral"""
|
||||
print("\n===== Test Mistral =====")
|
||||
|
||||
# Initialisation
|
||||
mistral = Mistral()
|
||||
mistral.prompt_system = "Tu es un assistant expert concis et précis. Réponds en français."
|
||||
|
||||
# Interrogation
|
||||
question = "Explique le concept de programmation orientée objet en 3 points clés."
|
||||
print(f"Question: {question}")
|
||||
response = mistral.Interroger(question)
|
||||
print(f"Réponse: {response}")
|
||||
|
||||
# Information sur la durée
|
||||
print(f"Temps de traitement: {mistral.dureeTraitement.total_seconds()} secondes")
|
||||
|
||||
def test_ollama():
|
||||
"""Test de la classe Ollama"""
|
||||
print("\n===== Test Ollama =====")
|
||||
|
||||
# Initialisation
|
||||
ollama = Ollama()
|
||||
ollama.prompt_system = "Tu es un assistant technique expert. Réponds en français."
|
||||
|
||||
# Configuration des paramètres avancés
|
||||
ollama.o_temperature = 0.7
|
||||
ollama.o_num_ctx = 4096
|
||||
|
||||
# Interrogation
|
||||
question = "Quelles sont les différences entre Python et JavaScript?"
|
||||
print(f"Question: {question}")
|
||||
response = ollama.Interroger(question)
|
||||
print(f"Réponse: {response}")
|
||||
|
||||
def test_deepseek():
|
||||
"""Test de la classe DeepSeek"""
|
||||
print("\n===== Test DeepSeek =====")
|
||||
|
||||
# Initialisation
|
||||
deepseek = DeepSeek()
|
||||
deepseek.prompt_system = "Tu es un assistant IA. Réponds en français."
|
||||
|
||||
# Configuration
|
||||
deepseek.max_tokens = 1000
|
||||
|
||||
# Interrogation
|
||||
question = "Explique-moi le théorème de Bayes de manière simple."
|
||||
print(f"Question: {question}")
|
||||
response = deepseek.Interroger(question)
|
||||
print(f"Réponse: {response}")
|
||||
|
||||
def test_perplexity():
|
||||
"""Test de la classe Perplexity"""
|
||||
print("\n===== Test Perplexity =====")
|
||||
|
||||
# Initialisation
|
||||
perplexity = Perplexity()
|
||||
# Le prompt system est déjà défini par défaut
|
||||
|
||||
# Configuration de recherche
|
||||
perplexity.search_recency_filter = "year"
|
||||
|
||||
# Interrogation
|
||||
question = "Quelles sont les dernières avancées en intelligence artificielle en 2023?"
|
||||
print(f"Question: {question}")
|
||||
response = perplexity.Interroger(question)
|
||||
print(f"Réponse: {response}")
|
||||
|
||||
def test_deepl():
|
||||
"""Test de la classe DeepL pour la traduction"""
|
||||
print("\n===== Test DeepL =====")
|
||||
|
||||
# Initialisation
|
||||
translator = DeepL()
|
||||
|
||||
# Traduction français -> anglais
|
||||
translator.langueSource = "FR"
|
||||
translator.langueDesti = "EN"
|
||||
text = "L'intelligence artificielle révolutionne notre façon de travailler et de vivre."
|
||||
print(f"Texte original (FR): {text}")
|
||||
translated = translator.Interroger(text)
|
||||
print(f"Traduction (EN): {translated}")
|
||||
|
||||
# Traduction anglais -> français
|
||||
translator.langueSource = "EN"
|
||||
translator.langueDesti = "FR"
|
||||
text = "Artificial intelligence is transforming the way we process and understand information."
|
||||
print(f"Texte original (EN): {text}")
|
||||
translated = translator.Interroger(text)
|
||||
print(f"Traduction (FR): {translated}")
|
||||
|
||||
def test_rag():
|
||||
"""Test de la classe RAG"""
|
||||
print("\n===== Test RAG =====")
|
||||
|
||||
# Initialisation
|
||||
rag = RAG()
|
||||
|
||||
# Interrogation
|
||||
question = "Quels sont les principes fondamentaux de l'apprentissage par renforcement?"
|
||||
print(f"Question formatée: {RAG.formateQuestion(question)}")
|
||||
|
||||
# Note: Cette partie peut échouer si le service RAG n'est pas accessible
|
||||
try:
|
||||
response = rag.Chat(question)
|
||||
print(f"Réponse: {response}")
|
||||
except Exception as e:
|
||||
print(f"Erreur lors de l'interrogation RAG: {e}")
|
||||
|
||||
def main():
|
||||
"""Fonction principale qui exécute tous les tests"""
|
||||
print("TESTS DES CLASSES LLM")
|
||||
print("=====================")
|
||||
|
||||
# Sélectionnez les tests à exécuter en décommentant les lignes
|
||||
test_mistral()
|
||||
test_ollama()
|
||||
test_deepseek()
|
||||
test_perplexity()
|
||||
test_deepl()
|
||||
test_rag()
|
||||
|
||||
print("\nTous les tests sont terminés!")
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
189
llm_classes/example_vision.py
Normal file
189
llm_classes/example_vision.py
Normal file
@ -0,0 +1,189 @@
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
"""
|
||||
Script d'exemple pour tester LlamaVision et les agents d'analyse d'images
|
||||
"""
|
||||
|
||||
import os
|
||||
import json
|
||||
import argparse
|
||||
from datetime import datetime
|
||||
from llm_classes import LlamaVision, AgentAnalyseImage, AgentAnalyseJSON, AgentQuestionReponse
|
||||
|
||||
def create_output_dir():
|
||||
"""Crée un répertoire pour les sorties basé sur la date/heure"""
|
||||
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||
output_dir = f"output_vision_{timestamp}"
|
||||
os.makedirs(output_dir, exist_ok=True)
|
||||
return output_dir
|
||||
|
||||
def analyse_directe(image_path, output_dir):
|
||||
"""Analyse directe avec LlamaVision sans passer par un agent"""
|
||||
print(f"\n=== Analyse directe de l'image {os.path.basename(image_path)} ===")
|
||||
|
||||
# Création d'une instance LlamaVision
|
||||
vision = LlamaVision()
|
||||
vision.prompt_system = "Tu es un assistant d'analyse d'images très précis. Décris l'image en détail."
|
||||
|
||||
# Chargement de l'image
|
||||
if not vision.set_image(image_path):
|
||||
print(f"Erreur: Impossible de charger l'image {image_path}")
|
||||
return
|
||||
|
||||
# Analyse de l'image
|
||||
question = "Décris en détail cette image. Qu'est-ce que tu y vois? Y a-t-il du texte visible?"
|
||||
print(f"Question: {question}")
|
||||
|
||||
reponse = vision.Interroger(question)
|
||||
print("\nRéponse:")
|
||||
print(reponse)
|
||||
|
||||
# Sauvegarde du résultat
|
||||
output_file = os.path.join(output_dir, "resultat_direct.json")
|
||||
vision.sauvegarder_resultats(output_file)
|
||||
print(f"\nRésultat sauvegardé dans: {output_file}")
|
||||
|
||||
def workflow_complet(image_path, json_path=None, output_dir=None):
|
||||
"""
|
||||
Workflow complet d'analyse d'image et traitement des données
|
||||
|
||||
Args:
|
||||
image_path: Chemin vers l'image à analyser
|
||||
json_path: Chemin vers un fichier JSON optionnel contenant des données additionnelles
|
||||
output_dir: Répertoire de sortie (si None, en crée un nouveau)
|
||||
"""
|
||||
if output_dir is None:
|
||||
output_dir = create_output_dir()
|
||||
|
||||
print(f"\n=== Workflow complet pour l'image {os.path.basename(image_path)} ===")
|
||||
print(f"Résultats sauvegardés dans: {output_dir}")
|
||||
|
||||
# Étape 1: Chargement des données JSON existantes si disponibles
|
||||
json_data = None
|
||||
if json_path and os.path.exists(json_path):
|
||||
try:
|
||||
with open(json_path, 'r', encoding='utf-8') as f:
|
||||
json_data = json.load(f)
|
||||
print(f"Données JSON chargées depuis: {json_path}")
|
||||
except Exception as e:
|
||||
print(f"Erreur lors du chargement du JSON: {e}")
|
||||
|
||||
# Étape 2: Analyse de l'image avec l'agent
|
||||
print("\n--- Étape 1: Analyse de l'image ---")
|
||||
agent_image = AgentAnalyseImage("Agent Vision")
|
||||
|
||||
# Configuration des paramètres
|
||||
agent_image.ajuster_parametres(temperature=0.7, top_p=0.9, num_ctx=4096)
|
||||
|
||||
# Analyse de l'image
|
||||
resultats_analyse = agent_image.executer(image_path, json_data)
|
||||
|
||||
# Sauvegarde des résultats intermédiaires
|
||||
output_image = os.path.join(output_dir, "analyse_image.json")
|
||||
agent_image.sauvegarder_resultats(output_image, resultats_analyse)
|
||||
print(f"Analyse d'image sauvegardée dans: {output_image}")
|
||||
|
||||
# Étape 3: Analyse des données JSON avec l'agent
|
||||
print("\n--- Étape 2: Analyse des données JSON ---")
|
||||
agent_json = AgentAnalyseJSON("Agent JSON")
|
||||
|
||||
# Extraction de la structure
|
||||
structure = agent_json.extraire_structure(resultats_analyse)
|
||||
|
||||
# Analyse des données
|
||||
question_analyse = "Analyse ces données extraites de l'image et résume les points principaux. Identifie les éléments clés et leur importance."
|
||||
resultats_json = agent_json.executer(resultats_analyse, question_analyse)
|
||||
|
||||
# Fusion des résultats
|
||||
resultats_complets = agent_json.fusionner_jsons(
|
||||
resultats_analyse,
|
||||
{
|
||||
"meta_analyse": {
|
||||
"timestamp": datetime.now().isoformat(),
|
||||
"structure": structure,
|
||||
"analyse_texte": resultats_json
|
||||
}
|
||||
}
|
||||
)
|
||||
|
||||
# Sauvegarde des résultats intermédiaires
|
||||
output_json = os.path.join(output_dir, "analyse_complete.json")
|
||||
with open(output_json, 'w', encoding='utf-8') as f:
|
||||
json.dump(resultats_complets, f, ensure_ascii=False, indent=2)
|
||||
print(f"Analyse complète sauvegardée dans: {output_json}")
|
||||
|
||||
# Étape 4: Questions/Réponses sur les données
|
||||
print("\n--- Étape 3: Tests de questions/réponses ---")
|
||||
agent_qr = AgentQuestionReponse("Agent QR")
|
||||
|
||||
# Génération automatique de questions
|
||||
questions_auto = agent_qr.generer_questions(resultats_complets, nb_questions=3)
|
||||
print("\nQuestions générées automatiquement:")
|
||||
for i, q in enumerate(questions_auto):
|
||||
print(f"{i+1}. {q}")
|
||||
|
||||
# Questions standard + questions générées
|
||||
toutes_questions = agent_qr.questions_standard + questions_auto
|
||||
|
||||
# Exécution des questions/réponses
|
||||
resultats_qr = agent_qr.executer(resultats_complets, toutes_questions)
|
||||
|
||||
# Évaluation des réponses
|
||||
evaluation = agent_qr.evaluer_reponses(resultats_qr["resultats_qr"])
|
||||
|
||||
# Fusion et sauvegarde des résultats finaux
|
||||
resultats_finaux = {
|
||||
"image": os.path.basename(image_path),
|
||||
"date_analyse": datetime.now().isoformat(),
|
||||
"donnees": resultats_complets,
|
||||
"questions_reponses": resultats_qr,
|
||||
"evaluation": evaluation
|
||||
}
|
||||
|
||||
output_final = os.path.join(output_dir, "resultats_finaux.json")
|
||||
with open(output_final, 'w', encoding='utf-8') as f:
|
||||
json.dump(resultats_finaux, f, ensure_ascii=False, indent=2)
|
||||
print(f"\nRésultats finaux sauvegardés dans: {output_final}")
|
||||
|
||||
# Affichage des résultats
|
||||
print("\n=== Résumé des résultats ===")
|
||||
print(f"Note moyenne des réponses: {evaluation['meta']['note_moyenne']:.1f}/10")
|
||||
|
||||
# Affichage d'un exemple de question/réponse
|
||||
if resultats_qr["resultats_qr"]:
|
||||
exemple = resultats_qr["resultats_qr"][0]
|
||||
print(f"\nExemple de Q/R:")
|
||||
print(f"Q: {exemple['question']}")
|
||||
print(f"R: {exemple['reponse'][:200]}...")
|
||||
|
||||
return resultats_finaux
|
||||
|
||||
def main():
|
||||
"""Fonction principale"""
|
||||
parser = argparse.ArgumentParser(description="Test de LlamaVision et des agents d'analyse d'images")
|
||||
parser.add_argument("--image", type=str, help="Chemin vers l'image à analyser", required=True)
|
||||
parser.add_argument("--json", type=str, help="Chemin vers un fichier JSON optionnel avec des données additionnelles")
|
||||
parser.add_argument("--mode", type=str, choices=["direct", "complet"], default="complet",
|
||||
help="Mode d'analyse: direct (LlamaVision uniquement) ou complet (workflow avec agents)")
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
# Vérification de l'existence de l'image
|
||||
if not os.path.exists(args.image):
|
||||
print(f"Erreur: L'image {args.image} n'existe pas")
|
||||
return
|
||||
|
||||
# Création du répertoire de sortie
|
||||
output_dir = create_output_dir()
|
||||
|
||||
# Mode d'analyse
|
||||
if args.mode == "direct":
|
||||
analyse_directe(args.image, output_dir)
|
||||
else:
|
||||
workflow_complet(args.image, args.json, output_dir)
|
||||
|
||||
print("\nAnalyse terminée!")
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
216
llm_classes/llama_vision.py
Normal file
216
llm_classes/llama_vision.py
Normal file
@ -0,0 +1,216 @@
|
||||
import json
|
||||
import base64
|
||||
import os
|
||||
from typing import Dict, List, Any, Optional, Union, Tuple
|
||||
from datetime import timedelta
|
||||
import requests
|
||||
from .llm import LLM
|
||||
|
||||
class LlamaVision(LLM):
|
||||
"""
|
||||
Classe pour l'intégration avec le modèle Llama Vision 3.2 90B
|
||||
Cette classe hérite de la classe de base LLM et ajoute le support des images
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""
|
||||
Initialisation des attributs spécifiques à Llama Vision
|
||||
"""
|
||||
super().__init__()
|
||||
|
||||
# Configuration du modèle
|
||||
self.Modele = "llama-vision:3.2-90b"
|
||||
|
||||
# Attributs spécifiques pour Llama Vision
|
||||
self.image_path: str = ""
|
||||
self.image_data: Optional[bytes] = None
|
||||
self.json_data: Dict[str, Any] = {}
|
||||
|
||||
# Paramètres de génération
|
||||
self.o_mirostat: int = 0
|
||||
self.o_mirostat_eta: float = 0.1
|
||||
self.o_mirostat_tau: float = 5
|
||||
self.o_num_ctx: int = 4096 # Contexte plus large pour les images + texte
|
||||
self.o_repeat_last_n: int = 64
|
||||
self.o_repeat_penalty: float = 1.1
|
||||
self.o_temperature: float = 0.7 # Valeur par défaut adaptée pour les descriptions d'images
|
||||
self.o_seed: int = 0
|
||||
self.o_stop: List[str] = []
|
||||
self.o_num_predict: int = 1024 # Nombre de tokens à prédire
|
||||
self.o_top_k: int = 40
|
||||
self.o_top_p: float = 0.9
|
||||
self.o_min_p: float = 0
|
||||
|
||||
# Options supplémentaires
|
||||
self.keep_alive: timedelta = timedelta(minutes=5)
|
||||
self.raw: bool = False
|
||||
self._stream: bool = False
|
||||
|
||||
# Stockage des résultats
|
||||
self.dernier_resultat: Dict[str, Any] = {}
|
||||
self.historique_resultats: List[Dict[str, Any]] = []
|
||||
|
||||
def urlBase(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de base de l'API Ollama pour Llama Vision
|
||||
"""
|
||||
return "http:/217.182.105.173/:11434/"
|
||||
|
||||
def cleAPI(self) -> str:
|
||||
"""
|
||||
Retourne la clé API (vide pour Ollama local)
|
||||
"""
|
||||
return ""
|
||||
|
||||
def set_image(self, image_path: str) -> bool:
|
||||
"""
|
||||
Définit l'image à analyser à partir d'un chemin de fichier
|
||||
Retourne True si l'image a été chargée avec succès
|
||||
"""
|
||||
if not os.path.exists(image_path):
|
||||
return False
|
||||
|
||||
try:
|
||||
with open(image_path, "rb") as f:
|
||||
self.image_data = f.read()
|
||||
self.image_path = image_path
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"Erreur lors du chargement de l'image: {e}")
|
||||
return False
|
||||
|
||||
def set_image_data(self, image_data: bytes) -> bool:
|
||||
"""
|
||||
Définit directement les données de l'image à analyser
|
||||
"""
|
||||
try:
|
||||
self.image_data = image_data
|
||||
self.image_path = ""
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"Erreur lors de la définition des données d'image: {e}")
|
||||
return False
|
||||
|
||||
def set_json_data(self, json_data: Dict[str, Any]) -> bool:
|
||||
"""
|
||||
Définit les données JSON à associer à l'image
|
||||
"""
|
||||
try:
|
||||
self.json_data = json_data
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"Erreur lors de la définition des données JSON: {e}")
|
||||
return False
|
||||
|
||||
def Interroger(self, question: str) -> str:
|
||||
"""
|
||||
Interroge le modèle Llama Vision avec une question et l'image chargée
|
||||
"""
|
||||
if not self.image_data:
|
||||
return "Erreur: Aucune image n'a été chargée. Utilisez set_image() ou set_image_data() d'abord."
|
||||
|
||||
return self.LLM_POST(question)
|
||||
|
||||
def _even_LLM_POST(self, question: str) -> None:
|
||||
"""
|
||||
Préparation du contenu de la requête pour Llama Vision
|
||||
"""
|
||||
# Paramètres de base
|
||||
self._Contenu["model"] = self.Modele
|
||||
self._Contenu["system"] = self.prompt_system if self.prompt_system else "Tu es un assistant d'analyse d'images qui peut analyser des images et répondre à des questions en français. Sois précis et détaillé dans tes descriptions."
|
||||
self._Contenu["prompt"] = question
|
||||
self._Contenu["stream"] = self._stream
|
||||
self._Contenu["raw"] = self.raw
|
||||
|
||||
# Ajout de l'image en base64
|
||||
if self.image_data:
|
||||
base64_image = base64.b64encode(self.image_data).decode('utf-8')
|
||||
self._Contenu["images"] = [base64_image]
|
||||
|
||||
# Conversion de timedelta en nombre de secondes pour keep_alive
|
||||
if isinstance(self.keep_alive, timedelta):
|
||||
self._Contenu["keep_alive"] = int(self.keep_alive.total_seconds())
|
||||
|
||||
# Ajout des données JSON dans le prompt si disponibles
|
||||
if self.json_data:
|
||||
json_str = json.dumps(self.json_data, ensure_ascii=False, indent=2)
|
||||
# On ajoute le JSON au prompt pour qu'il soit traité avec l'image
|
||||
self._Contenu["prompt"] = f"{question}\n\nVoici des données JSON associées à cette image:\n```json\n{json_str}\n```"
|
||||
|
||||
# Options avancées pour la génération
|
||||
self._Contenu["options"] = {
|
||||
"mirostat": self.o_mirostat,
|
||||
"mirostat_eta": self.o_mirostat_eta,
|
||||
"mirostat_tau": self.o_mirostat_tau,
|
||||
"num_ctx": self.o_num_ctx,
|
||||
"repeat_last_n": self.o_repeat_last_n,
|
||||
"repeat_penalty": self.o_repeat_penalty,
|
||||
"temperature": self.o_temperature,
|
||||
"seed": self.o_seed,
|
||||
"stop": self.o_stop,
|
||||
"num_predict": self.o_num_predict,
|
||||
"top_k": self.o_top_k,
|
||||
"top_p": self.o_top_p,
|
||||
"min_p": self.o_min_p
|
||||
}
|
||||
|
||||
def _interrogerRetourneReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Extraction de la réponse à partir du JSON retourné par Llama Vision
|
||||
"""
|
||||
try:
|
||||
data = json.loads(reponse)
|
||||
self.dernier_resultat = data
|
||||
self.historique_resultats.append(data)
|
||||
|
||||
# Ollama retourne généralement la réponse dans la clé "response"
|
||||
if "response" in data:
|
||||
return data["response"]
|
||||
else:
|
||||
return str(data)
|
||||
except Exception as e:
|
||||
return f"Erreur lors de l'analyse de la réponse: {e}\n\nRéponse brute: {reponse}"
|
||||
|
||||
def fusionner_json_avec_resultats(self) -> Dict[str, Any]:
|
||||
"""
|
||||
Fusionne les données JSON existantes avec les résultats de l'analyse d'image
|
||||
"""
|
||||
if not self.dernier_resultat:
|
||||
return self.json_data
|
||||
|
||||
# Créer une copie du JSON original
|
||||
resultat_fusionne = self.json_data.copy() if self.json_data else {}
|
||||
|
||||
# Ajouter le résultat de l'analyse d'image
|
||||
if "analyse_image" not in resultat_fusionne:
|
||||
resultat_fusionne["analyse_image"] = []
|
||||
|
||||
# Ajouter le résultat à la liste des analyses
|
||||
nouvelle_analyse = {
|
||||
"modele": self.Modele,
|
||||
"reponse": self.dernier_resultat.get("response", ""),
|
||||
"parametres": {
|
||||
"temperature": self.o_temperature,
|
||||
"top_p": self.o_top_p,
|
||||
"num_ctx": self.o_num_ctx
|
||||
}
|
||||
}
|
||||
|
||||
resultat_fusionne["analyse_image"].append(nouvelle_analyse)
|
||||
|
||||
return resultat_fusionne
|
||||
|
||||
def sauvegarder_resultats(self, chemin_fichier: str) -> bool:
|
||||
"""
|
||||
Sauvegarde les résultats fusionnés dans un fichier JSON
|
||||
"""
|
||||
try:
|
||||
resultats_fusionnes = self.fusionner_json_avec_resultats()
|
||||
|
||||
with open(chemin_fichier, 'w', encoding='utf-8') as f:
|
||||
json.dump(resultats_fusionnes, f, ensure_ascii=False, indent=2)
|
||||
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"Erreur lors de la sauvegarde des résultats: {e}")
|
||||
return False
|
||||
149
llm_classes/llm.py
Normal file
149
llm_classes/llm.py
Normal file
@ -0,0 +1,149 @@
|
||||
import abc
|
||||
import json
|
||||
import time
|
||||
import requests
|
||||
from datetime import datetime, timedelta
|
||||
from typing import Dict, List, Any, Optional, Union
|
||||
|
||||
class LLM(abc.ABC):
|
||||
"""Classe abstraite de base pour tous les modèles LLM"""
|
||||
|
||||
def __init__(self):
|
||||
"""Initialisation des attributs communs"""
|
||||
# Attributs publics
|
||||
self.Modele: str = ""
|
||||
self.prompt_system: str = ""
|
||||
self.o_temperature: float = 0.8
|
||||
self.o_top_k: int = 40
|
||||
self.o_top_p: float = 0.9
|
||||
|
||||
# Constantes publiques
|
||||
self.dureeTraitement: timedelta = timedelta()
|
||||
self.reponseErreur: bool = False
|
||||
|
||||
# Attributs protégés
|
||||
self._Contenu: Dict[str, Any] = {}
|
||||
self._m_sUrlFonction: str = ""
|
||||
self._stream: bool = False
|
||||
self._m_sFormat: str = ""
|
||||
self._heureDepart: Optional[datetime] = None
|
||||
self._heureFin: Optional[datetime] = None
|
||||
|
||||
# Initialisation par défaut
|
||||
self._stream = False
|
||||
|
||||
@property
|
||||
def urlFonction(self) -> str:
|
||||
"""Getter pour l'URL de la fonction"""
|
||||
return self._m_sUrlFonction
|
||||
|
||||
@urlFonction.setter
|
||||
def urlFonction(self, valeur: str):
|
||||
"""Setter pour l'URL de la fonction"""
|
||||
self._m_sUrlFonction = valeur
|
||||
|
||||
@property
|
||||
def format(self) -> str:
|
||||
"""Getter pour le format"""
|
||||
return self._m_sFormat
|
||||
|
||||
@format.setter
|
||||
def format(self, valeur: str):
|
||||
"""Setter pour le format"""
|
||||
self._m_sFormat = valeur
|
||||
|
||||
def model_list(self) -> List[str]:
|
||||
"""Retourne la liste des modèles disponibles"""
|
||||
return []
|
||||
|
||||
def commandeAutorisation(self) -> str:
|
||||
"""Retourne la commande d'autorisation à utiliser pour les API"""
|
||||
return "Bearer"
|
||||
|
||||
def Interroger(self, question: str) -> str:
|
||||
"""Interroge le LLM avec une question et retourne la réponse"""
|
||||
response = self.LLM_POST(question)
|
||||
return response
|
||||
|
||||
def LLM_POST(self, question: str) -> str:
|
||||
"""
|
||||
Méthode interne pour communiquer avec l'API du LLM
|
||||
Cette méthode gère les appels HTTP et le formattage de la réponse
|
||||
"""
|
||||
erreur = ""
|
||||
reponse = ""
|
||||
|
||||
# Préparation de la requête
|
||||
url = self.urlBase() + self.urlFonction
|
||||
headers = {
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
# Ajout de l'autorisation si nécessaire
|
||||
if self.cleAPI() != "":
|
||||
headers["Authorization"] = f"{self.commandeAutorisation()} {self.cleAPI()}"
|
||||
|
||||
# Préparation du contenu
|
||||
self._Contenu = {}
|
||||
if self.Modele != "":
|
||||
self._Contenu["model"] = self.Modele
|
||||
|
||||
# Préparation spécifique à la classe fille
|
||||
self._even_LLM_POST(question)
|
||||
|
||||
# Envoi de la requête
|
||||
self._heureDepart = datetime.now()
|
||||
|
||||
try:
|
||||
# Timeout de 2 minutes (équivalent à duréeNonRéponse = 2min)
|
||||
response = requests.post(
|
||||
url=url,
|
||||
headers=headers,
|
||||
json=self._Contenu,
|
||||
timeout=120
|
||||
)
|
||||
|
||||
self._heureFin = datetime.now()
|
||||
self.dureeTraitement = self._heureFin - self._heureDepart
|
||||
|
||||
# Traitement de la réponse
|
||||
if response.status_code in [200, 201]:
|
||||
reponse = self._interrogerRetourneReponse(response.text)
|
||||
self.reponseErreur = False
|
||||
return reponse
|
||||
else:
|
||||
erreur = response.text
|
||||
self.reponseErreur = True
|
||||
return erreur
|
||||
|
||||
except Exception as e:
|
||||
self._heureFin = datetime.now()
|
||||
self.dureeTraitement = self._heureFin - self._heureDepart
|
||||
self.reponseErreur = True
|
||||
return str(e)
|
||||
|
||||
@abc.abstractmethod
|
||||
def urlBase(self) -> str:
|
||||
"""Retourne l'URL de base de l'API (à implémenter dans les classes filles)"""
|
||||
pass
|
||||
|
||||
@abc.abstractmethod
|
||||
def cleAPI(self) -> str:
|
||||
"""Retourne la clé API à utiliser (à implémenter dans les classes filles)"""
|
||||
pass
|
||||
|
||||
@abc.abstractmethod
|
||||
def _even_LLM_POST(self, question: str) -> None:
|
||||
"""
|
||||
Méthode appelée lors de la préparation de la requête POST
|
||||
Permet de personnaliser le contenu pour chaque type de LLM
|
||||
"""
|
||||
pass
|
||||
|
||||
@abc.abstractmethod
|
||||
def _interrogerRetourneReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Méthode pour extraire la réponse du LLM à partir du JSON retourné
|
||||
À implémenter selon le format de réponse spécifique à chaque LLM
|
||||
"""
|
||||
pass
|
||||
124
llm_classes/mistral.py
Normal file
124
llm_classes/mistral.py
Normal file
@ -0,0 +1,124 @@
|
||||
import json
|
||||
from typing import Dict, List, Any, Optional
|
||||
from .llm import LLM
|
||||
|
||||
class Mistral(LLM):
|
||||
"""
|
||||
Classe pour l'intégration avec l'API Mistral AI
|
||||
Cette classe hérite de la classe de base LLM
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""
|
||||
Initialisation des attributs spécifiques à Mistral
|
||||
"""
|
||||
super().__init__()
|
||||
|
||||
# Attributs spécifiques à Mistral
|
||||
self.maxToken: int = 1000
|
||||
self.seed: int = 0
|
||||
self.presence_penalty: float = 0
|
||||
self.frequency_penalty: float = 0
|
||||
self.n: int = 1
|
||||
self.prediction: Dict[str, str] = {"type": "content", "content": ""}
|
||||
self.safe_prompt: bool = False
|
||||
self.o_stop: str = "string"
|
||||
self._m_tabModels: List[str] = []
|
||||
|
||||
# Initialisation par défaut
|
||||
self.Modele = "mistral-large-latest"
|
||||
self.o_temperature = 0.2
|
||||
self.o_top_p = 1
|
||||
self.maxToken = 1000
|
||||
self.presence_penalty = 0
|
||||
self.frequency_penalty = 0
|
||||
self.n = 1
|
||||
self.prediction["type"] = "content"
|
||||
self.prediction["content"] = ""
|
||||
self.safe_prompt = False
|
||||
self.o_stop = "string"
|
||||
|
||||
def urlBase(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de base de l'API Mistral
|
||||
"""
|
||||
return "https://api.mistral.ai/v1/"
|
||||
|
||||
def cleAPI(self) -> str:
|
||||
"""
|
||||
Retourne la clé API Mistral
|
||||
"""
|
||||
return "2iGzTzE9csRQ9IoASoUjplHwEjA200Vh"
|
||||
|
||||
def urlFonction(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de la fonction par défaut
|
||||
"""
|
||||
return "chat/completions"
|
||||
|
||||
def model_list(self) -> List[str]:
|
||||
"""
|
||||
Récupère la liste des modèles disponibles sur Mistral AI
|
||||
Si la liste est déjà récupérée, retourne la liste en cache
|
||||
"""
|
||||
import requests
|
||||
|
||||
if len(self._m_tabModels) > 1:
|
||||
return self._m_tabModels
|
||||
|
||||
url = self.urlBase() + "models"
|
||||
headers = {
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
if self.cleAPI() != "":
|
||||
headers["Authorization"] = "Bearer " + self.cleAPI()
|
||||
|
||||
try:
|
||||
response = requests.get(url=url, headers=headers)
|
||||
|
||||
if response.status_code == 200:
|
||||
data = json.loads(response.text)
|
||||
for item in data["data"]:
|
||||
self._m_tabModels.append(item["id"])
|
||||
else:
|
||||
self._m_tabModels.append(response.text)
|
||||
except Exception as e:
|
||||
self._m_tabModels.append(str(e))
|
||||
|
||||
return self._m_tabModels
|
||||
|
||||
def _even_LLM_POST(self, question: str) -> None:
|
||||
"""
|
||||
Préparation du contenu de la requête pour Mistral
|
||||
"""
|
||||
self._Contenu["messages"] = [
|
||||
{"role": "system", "content": self.prompt_system},
|
||||
{"role": "user", "content": question}
|
||||
]
|
||||
|
||||
self._Contenu["temperature"] = self.o_temperature
|
||||
self._Contenu["top_p"] = self.o_top_p
|
||||
|
||||
if self.maxToken != 0:
|
||||
self._Contenu["max_tokens"] = self.maxToken
|
||||
|
||||
if self.seed != 0:
|
||||
self._Contenu["random_seed"] = self.seed
|
||||
|
||||
if self.format != "":
|
||||
self._Contenu["response_format"] = self.format
|
||||
|
||||
# Paramètres supplémentaires
|
||||
self._Contenu["presence_penalty"] = self.presence_penalty
|
||||
self._Contenu["frequency_penalty"] = self.frequency_penalty
|
||||
self._Contenu["n"] = 1
|
||||
self._Contenu["prediction"] = self.prediction
|
||||
self._Contenu["stop"] = self.o_stop
|
||||
|
||||
def _interrogerRetourneReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Extraction de la réponse à partir du JSON retourné par Mistral
|
||||
"""
|
||||
data = json.loads(reponse)
|
||||
return data["choices"][0]["message"]["content"]
|
||||
115
llm_classes/ollama.py
Normal file
115
llm_classes/ollama.py
Normal file
@ -0,0 +1,115 @@
|
||||
import json
|
||||
from typing import Dict, List, Any, Optional, Union
|
||||
from datetime import timedelta
|
||||
from .llm import LLM
|
||||
|
||||
class Ollama(LLM):
|
||||
"""
|
||||
Classe pour l'intégration avec l'API Ollama
|
||||
Cette classe hérite de la classe de base LLM
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""
|
||||
Initialisation des attributs spécifiques à Ollama
|
||||
"""
|
||||
super().__init__()
|
||||
|
||||
# Attributs spécifiques à Ollama
|
||||
self.suffix: str = ""
|
||||
self.images: bytes = b"" # Équivalent à Buffer en WLangage
|
||||
self.template: str = ""
|
||||
self.raw: bool = False
|
||||
self.keep_alive: timedelta = timedelta(minutes=5) # 5min en WLangage
|
||||
self.o_mirostat: int = 0
|
||||
self.o_mirostat_eta: float = 0.1
|
||||
self.o_mirostat_tau: float = 5
|
||||
self.o_num_ctx: int = 2048
|
||||
self.o_repeat_last_n: int = 64
|
||||
self.o_repeat_penalty: float = 1.1
|
||||
self.o_seed: int = 0
|
||||
self.o_stop: List[str] = []
|
||||
self.o_num_predict: int = -1
|
||||
self.o_min_p: float = 0
|
||||
|
||||
# Initialisation par défaut
|
||||
self.o_mirostat = 0
|
||||
self.o_mirostat_eta = 0.1
|
||||
self.o_mirostat_tau = 5
|
||||
self.o_num_ctx = 2048
|
||||
self.o_repeat_last_n = 64
|
||||
self.o_repeat_penalty = 1.1
|
||||
self.o_temperature = 0.8
|
||||
self.o_seed = 0
|
||||
self.o_stop = []
|
||||
self.o_num_predict = -1
|
||||
self.o_top_k = 40
|
||||
self.o_top_p = 0.9
|
||||
self.o_min_p = 0
|
||||
self.raw = False
|
||||
self.keep_alive = timedelta(minutes=5) # 5min en WLangage
|
||||
|
||||
def urlBase(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de base de l'API Ollama
|
||||
"""
|
||||
return "http://217.182.105.173:11434/"
|
||||
|
||||
def cleAPI(self) -> str:
|
||||
"""
|
||||
Retourne la clé API Ollama (vide par défaut)
|
||||
"""
|
||||
return ""
|
||||
|
||||
def Interroger(self, question: str) -> str:
|
||||
"""
|
||||
Interroge Ollama avec une question
|
||||
Modifie l'URL de fonction avant l'appel
|
||||
"""
|
||||
self.urlFonction = "api/generate"
|
||||
return self.LLM_POST(question)
|
||||
|
||||
def _even_LLM_POST(self, question: str) -> None:
|
||||
"""
|
||||
Préparation du contenu de la requête pour Ollama
|
||||
"""
|
||||
# Paramètres de base
|
||||
self._Contenu["system"] = self.prompt_system
|
||||
self._Contenu["prompt"] = question
|
||||
self._Contenu["stream"] = self._stream
|
||||
self._Contenu["suffix"] = self.suffix
|
||||
self._Contenu["format"] = self.format
|
||||
self._Contenu["raw"] = self.raw
|
||||
|
||||
# Conversion de timedelta en nombre de secondes pour keep_alive
|
||||
if isinstance(self.keep_alive, timedelta):
|
||||
self._Contenu["keep_alive"] = int(self.keep_alive.total_seconds())
|
||||
|
||||
# Traitement des images si présentes
|
||||
if self.images:
|
||||
# À implémenter si nécessaire
|
||||
pass
|
||||
|
||||
# Options avancées
|
||||
self._Contenu["options"] = {
|
||||
"mirostat": self.o_mirostat,
|
||||
"mirostat_eta": self.o_mirostat_eta,
|
||||
"mirostat_tau": self.o_mirostat_tau,
|
||||
"num_ctx": self.o_num_ctx,
|
||||
"repeat_last_n": self.o_repeat_last_n,
|
||||
"repeat_penalty": self.o_repeat_penalty,
|
||||
"temperature": self.o_temperature,
|
||||
"seed": self.o_seed,
|
||||
"stop": self.o_stop,
|
||||
"num_predict": self.o_num_predict,
|
||||
"top_k": self.o_top_k,
|
||||
"top_p": self.o_top_p,
|
||||
"min_p": self.o_min_p
|
||||
}
|
||||
|
||||
def _interrogerRetourneReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Extraction de la réponse à partir du JSON retourné par Ollama
|
||||
"""
|
||||
data = json.loads(reponse)
|
||||
return data["response"]
|
||||
99
llm_classes/perplexity.py
Normal file
99
llm_classes/perplexity.py
Normal file
@ -0,0 +1,99 @@
|
||||
import json
|
||||
from typing import Dict, List, Any, Optional, Union
|
||||
from .llm import LLM
|
||||
|
||||
class Perplexity(LLM):
|
||||
"""
|
||||
Classe pour l'intégration avec l'API Perplexity
|
||||
Cette classe hérite de la classe de base LLM
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""
|
||||
Initialisation des attributs spécifiques à Perplexity
|
||||
"""
|
||||
super().__init__()
|
||||
|
||||
# Attributs spécifiques à Perplexity
|
||||
self.maxToken: int = 1000
|
||||
self.presence_penalty: int = 0
|
||||
self.frequency_penaly: int = 1 # Note: typo dans le code source - penaly au lieu de penalty
|
||||
self.search_recency_filter: str = "all"
|
||||
self.search_domain_filter: List[str] = []
|
||||
|
||||
# Initialisation par défaut
|
||||
self.maxToken = 1000
|
||||
self.o_temperature = 0.5
|
||||
self.o_top_p = 0.8
|
||||
self.o_top_k = 0
|
||||
self.presence_penalty = 0
|
||||
self.frequency_penaly = 1
|
||||
self.search_recency_filter = "all" # "year" est une autre option
|
||||
self.prompt_system = "Soyez précis et concis. Répondez uniquement en français. Effectuez une recherche approfondie et fournissez des informations détaillées et à jour."
|
||||
|
||||
def cleAPI(self) -> str:
|
||||
"""
|
||||
Retourne la clé API Perplexity
|
||||
"""
|
||||
return "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"
|
||||
|
||||
def urlBase(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de base de l'API Perplexity
|
||||
"""
|
||||
return "https://api.perplexity.ai/chat/completions"
|
||||
|
||||
def Interroger(self, question: str) -> str:
|
||||
"""
|
||||
Interroge Perplexity avec une question
|
||||
Définit le modèle par défaut à "sonar"
|
||||
"""
|
||||
self.Modele = "sonar"
|
||||
return self.LLM_POST(question)
|
||||
|
||||
def _even_LLM_POST(self, question: str) -> None:
|
||||
"""
|
||||
Préparation du contenu de la requête pour Perplexity
|
||||
"""
|
||||
self._Contenu["messages"] = [
|
||||
{"role": "system", "content": self.prompt_system},
|
||||
{"role": "user", "content": question}
|
||||
]
|
||||
|
||||
# Ajout du nombre maximal de tokens si défini
|
||||
if self.maxToken != 0:
|
||||
self._Contenu["max_tokens"] = self.maxToken
|
||||
|
||||
# Température et format
|
||||
self._Contenu["temperature"] = self.o_temperature
|
||||
if self.format:
|
||||
self._Contenu["response_format"] = self.format
|
||||
|
||||
# Paramètres de filtrage
|
||||
self._Contenu["top_p"] = self.o_top_p
|
||||
|
||||
# Ajout des domaines de recherche si définis
|
||||
if len(self.search_domain_filter) > 0:
|
||||
# Exemple de filtrage: ["perplexity.ai"]
|
||||
self._Contenu["search_domain_filter"] = self.search_domain_filter
|
||||
|
||||
# Options supplémentaires
|
||||
self._Contenu["return_images"] = False
|
||||
self._Contenu["return_related_questions"] = False
|
||||
|
||||
# Filtre de récence si autre que "all"
|
||||
if self.search_recency_filter != "all":
|
||||
self._Contenu["search_recency_filter"] = self.search_recency_filter
|
||||
|
||||
# Autres paramètres
|
||||
self._Contenu["top_k"] = self.o_top_k
|
||||
self._Contenu["presence_penalty"] = self.presence_penalty
|
||||
self._Contenu["frequency_penalty"] = self.frequency_penaly
|
||||
self._Contenu["response_format"] = None # Null en WLangage
|
||||
|
||||
def _interrogerRetourneReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Extraction de la réponse à partir du JSON retourné par Perplexity
|
||||
"""
|
||||
data = json.loads(reponse)
|
||||
return data["choices"][0]["message"]["content"]
|
||||
183
llm_classes/rag.py
Normal file
183
llm_classes/rag.py
Normal file
@ -0,0 +1,183 @@
|
||||
import json
|
||||
import datetime
|
||||
import requests
|
||||
from typing import Dict, List, Any, Optional, Union, Tuple
|
||||
|
||||
class RAG:
|
||||
"""
|
||||
Classe pour l'intégration avec l'API RAG (Retrieval Augmented Generation)
|
||||
Cette classe ne dépend pas de la classe LLM et fonctionne de manière autonome
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""
|
||||
Initialisation des attributs de RAG
|
||||
"""
|
||||
# Attributs publics constants (transformés en attributs d'instance)
|
||||
self.sessionID: str = ""
|
||||
|
||||
# Attributs privés
|
||||
self._reqRAG: requests.Session = requests.Session()
|
||||
self._taContenu: Dict[str, Any] = {}
|
||||
self._m_sChatID: str = ""
|
||||
self._mn_Question: int = 0
|
||||
|
||||
# Initialisation par défaut
|
||||
self._reqRAG.headers["Authorization"] = "Bearer ragflow-c4YTNkMzcwZDM1ODExZWZiODA2MDI0Mm"
|
||||
self._reqRAG.headers["Content-Type"] = "application/json"
|
||||
|
||||
def _urlBase(self) -> str:
|
||||
"""
|
||||
Retourne l'URL de base de l'API RAG
|
||||
"""
|
||||
return "http://10.103.0.100/api/v1/chats/"
|
||||
|
||||
def _chatID(self) -> str:
|
||||
"""
|
||||
Retourne l'ID de chat actuel ou un ID par défaut
|
||||
"""
|
||||
if self._m_sChatID:
|
||||
return self._m_sChatID
|
||||
else:
|
||||
return "ffb1058ed4b811ef8a900242ac120003"
|
||||
|
||||
def _set_chatID(self, valeur: str) -> None:
|
||||
"""
|
||||
Définit l'ID de chat
|
||||
"""
|
||||
self._m_sChatID = valeur
|
||||
|
||||
def _Contenu(self) -> str:
|
||||
"""
|
||||
Sérialise le contenu en JSON
|
||||
"""
|
||||
return json.dumps(self._taContenu)
|
||||
|
||||
def _urlChat(self) -> str:
|
||||
"""
|
||||
Retourne l'URL complète pour le chat
|
||||
"""
|
||||
return f"{self._urlBase()}{self._chatID()}/"
|
||||
|
||||
def __del__(self):
|
||||
"""
|
||||
Destructeur - supprime la session de chat si nécessaire
|
||||
"""
|
||||
# Équivalent de: SI mn_Question = 0 OU EnModeTest() ALORS supprimeSessionChat()
|
||||
if self._mn_Question == 0:
|
||||
self._supprimeSessionChat()
|
||||
|
||||
def _creeSessionChat(self) -> Tuple[bool, str]:
|
||||
"""
|
||||
Crée une nouvelle session de chat
|
||||
Retourne un tuple (succès, message)
|
||||
"""
|
||||
self._taContenu.clear()
|
||||
|
||||
# Préparation de la requête - sans utiliser l'attribut url directement
|
||||
req_url = f"{self._urlChat()}sessions"
|
||||
|
||||
# Nom de la session basé sur la date et l'heure actuelles
|
||||
current_datetime = datetime.datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||
self._taContenu["name"] = current_datetime
|
||||
|
||||
try:
|
||||
response = self._reqRAG.post(
|
||||
url=req_url,
|
||||
json=self._taContenu
|
||||
)
|
||||
|
||||
data = json.loads(response.text)
|
||||
|
||||
if data.get("code") == 0:
|
||||
self.sessionID = data["data"]["id"]
|
||||
return True, self._formateReponse(data["data"]["messages"][0]["content"])
|
||||
else:
|
||||
return False, self._formateReponse(data.get("message", "Erreur inconnue"))
|
||||
|
||||
except Exception as e:
|
||||
return False, self._formateReponse(str(e))
|
||||
|
||||
def Chat(self, question: str) -> str:
|
||||
"""
|
||||
Envoie une question au chat RAG
|
||||
"""
|
||||
# Crée une session si nécessaire
|
||||
if not self.sessionID:
|
||||
success, response = self._creeSessionChat()
|
||||
if not success:
|
||||
return self._formateReponse(response)
|
||||
|
||||
# Incrémente le compteur de questions
|
||||
self._mn_Question += 1
|
||||
|
||||
# Prépare la requête
|
||||
self._taContenu.clear()
|
||||
self._taContenu["question"] = question
|
||||
self._taContenu["stream"] = False
|
||||
self._taContenu["session_id"] = self.sessionID
|
||||
|
||||
try:
|
||||
response = self._reqRAG.post(
|
||||
url=self._urlChat() + "completions",
|
||||
json=self._taContenu
|
||||
)
|
||||
|
||||
data = json.loads(response.text)
|
||||
|
||||
if data.get("code") == 0:
|
||||
return self._formateReponse(data["data"]["answer"])
|
||||
else:
|
||||
return self._formateReponse(data.get("message", "Erreur inconnue"))
|
||||
|
||||
except Exception as e:
|
||||
return self._formateReponse(str(e))
|
||||
|
||||
def _formateReponse(self, reponse: str) -> str:
|
||||
"""
|
||||
Formate la réponse en HTML
|
||||
"""
|
||||
# Conversion en HTML avec style personnalisé
|
||||
html = f"""<div style="background-color: #e8f4ff; border: 1px solid #c2e0ff; border-radius: 10px; padding: 10px 15px; font-family: Arial, sans-serif; font-size: 14px; color: #000; display: inline-block;">
|
||||
{reponse}
|
||||
</div>"""
|
||||
|
||||
return html
|
||||
|
||||
@staticmethod
|
||||
def formateQuestion(question: str) -> str:
|
||||
"""
|
||||
Formate la question en HTML
|
||||
"""
|
||||
# Conversion en HTML avec style personnalisé
|
||||
html = f"""<div style="text-align: right; font-family: Arial, sans-serif; font-size: 14px; color: #000; padding: 10px 15px;">
|
||||
{question}
|
||||
</div>"""
|
||||
|
||||
return html
|
||||
|
||||
def _supprimeSessionChat(self) -> str:
|
||||
"""
|
||||
Supprime la session de chat actuelle
|
||||
"""
|
||||
if not self.sessionID:
|
||||
return "Aucune session à supprimer"
|
||||
|
||||
self._taContenu.clear()
|
||||
self._taContenu["ids"] = [self.sessionID]
|
||||
|
||||
try:
|
||||
response = self._reqRAG.delete(
|
||||
url=self._urlChat() + "/sessions",
|
||||
json=self._taContenu
|
||||
)
|
||||
|
||||
data = json.loads(response.text)
|
||||
|
||||
if data.get("code") == 0:
|
||||
return "OK"
|
||||
else:
|
||||
return data.get("message", "Erreur inconnue")
|
||||
|
||||
except Exception as e:
|
||||
return str(e)
|
||||
7
llm_classes/requirements.txt
Normal file
7
llm_classes/requirements.txt
Normal file
@ -0,0 +1,7 @@
|
||||
requests>=2.28.0
|
||||
typing>=3.7.4.3
|
||||
datetime>=4.3
|
||||
json>=2.0.9
|
||||
argparse>=1.4.0
|
||||
pillow>=9.0.0 # Pour le traitement d'images
|
||||
base64>=1.0.0 # Pour l'encodage d'images
|
||||
28
prompt.txt
28
prompt.txt
@ -7,3 +7,31 @@ Pour chaques points ci-dessous je voudrai que tu considère que je suis un débu
|
||||
5. Création d'agent avec Ollama (utilisation de classes ollama général, puis classes modèle dédié)
|
||||
6. Optimisation rapide et pratique du paramétrage des paramètres du modèle ollama global et des autres modèles
|
||||
7. Optimisation et intégration dans vscode plus précisement j'utilise cursor (la aussi explication et mise en place détaillée)
|
||||
|
||||
|
||||
Tu es un spécialiste en WLangage plus précisement sur webdev:
|
||||
- Je dois récupérer du code WLanguage afin de le convertir en python
|
||||
- J'ai le projet qui qui ouvert devant mais je ne sais pas manipuler webdev
|
||||
- Je voudrai dans un premier temps que tu m'expliques comment se structure le code dans webdev
|
||||
- Je voudrai pouvoir extraire tout le code de ce projet:
|
||||
- dans un fichier adapté pour pouvoir par la suite l'adapter
|
||||
Tu peux te référer à la documentation PC soft en ligne
|
||||
Je voudrai que tu me guides pas à pas pour réaliser cette tâche
|
||||
Je veux que tu te focalise uniquement sur cette partie de récupération, nous verrons plus tard pour le rest
|
||||
je voudrai que tu me résumes ma demande avant de commencer afin de réajuter si besoin ma demande
|
||||
|
||||
J'ai recopié le code du projet robo_mat dans ce fichier .md j'y ai mis tous les éléments dans l'ordre:
|
||||
- Les sections vides signifie qu'il n'y a rien de précisement
|
||||
- Mes demandes:
|
||||
2. Je voudrai que le voit dans l'ordre (page, puis classes par classes) ce que tu comprends du code.
|
||||
- Je te rappelle que je ne connais pas webdev et que l'objectif est d'en apprendre plus sur son fonctionnement
|
||||
- Les classes font partie d'un composant interne qui s'appelle LLM qui comprend (Pages, Etats, Requêtes, Classes), pour le moment seul classes contient du code
|
||||
- je voudrai comprendre comment s'imbrique les sections et sous sections à gauche par exemple classes à des sous-sections (par ex( classe déclaré clsDeepl) qui a des sous sections(Membres, Propriétés, Constructeur etc...))
|
||||
3. Une la partie compréhension webdev terminée:
|
||||
- Je voudrai que tu mettes en parallèle les points communs de structures par rapport à python
|
||||
- Mettre en place un plan détaillé de l'adaptation en python
|
||||
-Mettre en place une adaptation en python partie de code par partie de code
|
||||
Formatage de la réponse:
|
||||
- je voudrai que ta réponse soit dans un fichier .md que tu pourras ajuster et corrigé en fonction de mes ajustements et qu'il soit téléchargeable
|
||||
- je te joint le fichier .md du code récupérer sur le projet
|
||||
|
||||
|
||||
13
setup.py
Normal file
13
setup.py
Normal file
@ -0,0 +1,13 @@
|
||||
from setuptools import setup, find_packages
|
||||
|
||||
setup(
|
||||
name="llm_classes",
|
||||
version="0.1",
|
||||
packages=find_packages(),
|
||||
install_requires=[
|
||||
"requests>=2.28.0",
|
||||
"datetime>=4.3",
|
||||
"pillow>=9.0.0",
|
||||
"argparse>=1.4.0"
|
||||
],
|
||||
)
|
||||
47
test.md
47
test.md
@ -1,47 +0,0 @@
|
||||
### 1. Connaissances de base et commandes essentielles
|
||||
|
||||
#### a. Installation et configuration initiale
|
||||
|
||||
- Installation sur Linux/Windows/macOS:
|
||||
|
||||
```bash
|
||||
curl -fsSL https://ollama.com/install.sh | sh
|
||||
```
|
||||
|
||||
- Vérifier l'installation :
|
||||
|
||||
```bash
|
||||
ollama --version
|
||||
```
|
||||
|
||||
- Lancer le serveur Ollama :
|
||||
|
||||
```bash
|
||||
ollama serve
|
||||
```
|
||||
#### Commandes complètes de base
|
||||
|
||||
- Afficher les modèles disponibles :
|
||||
|
||||
```bash
|
||||
ollama list
|
||||
```
|
||||
|
||||
- Télécharger un modèle précis :
|
||||
|
||||
```bash
|
||||
ollama pull llama3
|
||||
```
|
||||
|
||||
- Lancer un prompt rapidement :
|
||||
|
||||
```bash
|
||||
ollma run llama3 "Quelle est la capitale de la France?"
|
||||
```
|
||||
|
||||
- Gestion avancée des modèles :
|
||||
|
||||
```bash
|
||||
ollama show llama3
|
||||
ollama rm llama3
|
||||
```
|
||||
247
testclass/bin/Activate.ps1
Normal file
247
testclass/bin/Activate.ps1
Normal file
@ -0,0 +1,247 @@
|
||||
<#
|
||||
.Synopsis
|
||||
Activate a Python virtual environment for the current PowerShell session.
|
||||
|
||||
.Description
|
||||
Pushes the python executable for a virtual environment to the front of the
|
||||
$Env:PATH environment variable and sets the prompt to signify that you are
|
||||
in a Python virtual environment. Makes use of the command line switches as
|
||||
well as the `pyvenv.cfg` file values present in the virtual environment.
|
||||
|
||||
.Parameter VenvDir
|
||||
Path to the directory that contains the virtual environment to activate. The
|
||||
default value for this is the parent of the directory that the Activate.ps1
|
||||
script is located within.
|
||||
|
||||
.Parameter Prompt
|
||||
The prompt prefix to display when this virtual environment is activated. By
|
||||
default, this prompt is the name of the virtual environment folder (VenvDir)
|
||||
surrounded by parentheses and followed by a single space (ie. '(.venv) ').
|
||||
|
||||
.Example
|
||||
Activate.ps1
|
||||
Activates the Python virtual environment that contains the Activate.ps1 script.
|
||||
|
||||
.Example
|
||||
Activate.ps1 -Verbose
|
||||
Activates the Python virtual environment that contains the Activate.ps1 script,
|
||||
and shows extra information about the activation as it executes.
|
||||
|
||||
.Example
|
||||
Activate.ps1 -VenvDir C:\Users\MyUser\Common\.venv
|
||||
Activates the Python virtual environment located in the specified location.
|
||||
|
||||
.Example
|
||||
Activate.ps1 -Prompt "MyPython"
|
||||
Activates the Python virtual environment that contains the Activate.ps1 script,
|
||||
and prefixes the current prompt with the specified string (surrounded in
|
||||
parentheses) while the virtual environment is active.
|
||||
|
||||
.Notes
|
||||
On Windows, it may be required to enable this Activate.ps1 script by setting the
|
||||
execution policy for the user. You can do this by issuing the following PowerShell
|
||||
command:
|
||||
|
||||
PS C:\> Set-ExecutionPolicy -ExecutionPolicy RemoteSigned -Scope CurrentUser
|
||||
|
||||
For more information on Execution Policies:
|
||||
https://go.microsoft.com/fwlink/?LinkID=135170
|
||||
|
||||
#>
|
||||
Param(
|
||||
[Parameter(Mandatory = $false)]
|
||||
[String]
|
||||
$VenvDir,
|
||||
[Parameter(Mandatory = $false)]
|
||||
[String]
|
||||
$Prompt
|
||||
)
|
||||
|
||||
<# Function declarations --------------------------------------------------- #>
|
||||
|
||||
<#
|
||||
.Synopsis
|
||||
Remove all shell session elements added by the Activate script, including the
|
||||
addition of the virtual environment's Python executable from the beginning of
|
||||
the PATH variable.
|
||||
|
||||
.Parameter NonDestructive
|
||||
If present, do not remove this function from the global namespace for the
|
||||
session.
|
||||
|
||||
#>
|
||||
function global:deactivate ([switch]$NonDestructive) {
|
||||
# Revert to original values
|
||||
|
||||
# The prior prompt:
|
||||
if (Test-Path -Path Function:_OLD_VIRTUAL_PROMPT) {
|
||||
Copy-Item -Path Function:_OLD_VIRTUAL_PROMPT -Destination Function:prompt
|
||||
Remove-Item -Path Function:_OLD_VIRTUAL_PROMPT
|
||||
}
|
||||
|
||||
# The prior PYTHONHOME:
|
||||
if (Test-Path -Path Env:_OLD_VIRTUAL_PYTHONHOME) {
|
||||
Copy-Item -Path Env:_OLD_VIRTUAL_PYTHONHOME -Destination Env:PYTHONHOME
|
||||
Remove-Item -Path Env:_OLD_VIRTUAL_PYTHONHOME
|
||||
}
|
||||
|
||||
# The prior PATH:
|
||||
if (Test-Path -Path Env:_OLD_VIRTUAL_PATH) {
|
||||
Copy-Item -Path Env:_OLD_VIRTUAL_PATH -Destination Env:PATH
|
||||
Remove-Item -Path Env:_OLD_VIRTUAL_PATH
|
||||
}
|
||||
|
||||
# Just remove the VIRTUAL_ENV altogether:
|
||||
if (Test-Path -Path Env:VIRTUAL_ENV) {
|
||||
Remove-Item -Path env:VIRTUAL_ENV
|
||||
}
|
||||
|
||||
# Just remove VIRTUAL_ENV_PROMPT altogether.
|
||||
if (Test-Path -Path Env:VIRTUAL_ENV_PROMPT) {
|
||||
Remove-Item -Path env:VIRTUAL_ENV_PROMPT
|
||||
}
|
||||
|
||||
# Just remove the _PYTHON_VENV_PROMPT_PREFIX altogether:
|
||||
if (Get-Variable -Name "_PYTHON_VENV_PROMPT_PREFIX" -ErrorAction SilentlyContinue) {
|
||||
Remove-Variable -Name _PYTHON_VENV_PROMPT_PREFIX -Scope Global -Force
|
||||
}
|
||||
|
||||
# Leave deactivate function in the global namespace if requested:
|
||||
if (-not $NonDestructive) {
|
||||
Remove-Item -Path function:deactivate
|
||||
}
|
||||
}
|
||||
|
||||
<#
|
||||
.Description
|
||||
Get-PyVenvConfig parses the values from the pyvenv.cfg file located in the
|
||||
given folder, and returns them in a map.
|
||||
|
||||
For each line in the pyvenv.cfg file, if that line can be parsed into exactly
|
||||
two strings separated by `=` (with any amount of whitespace surrounding the =)
|
||||
then it is considered a `key = value` line. The left hand string is the key,
|
||||
the right hand is the value.
|
||||
|
||||
If the value starts with a `'` or a `"` then the first and last character is
|
||||
stripped from the value before being captured.
|
||||
|
||||
.Parameter ConfigDir
|
||||
Path to the directory that contains the `pyvenv.cfg` file.
|
||||
#>
|
||||
function Get-PyVenvConfig(
|
||||
[String]
|
||||
$ConfigDir
|
||||
) {
|
||||
Write-Verbose "Given ConfigDir=$ConfigDir, obtain values in pyvenv.cfg"
|
||||
|
||||
# Ensure the file exists, and issue a warning if it doesn't (but still allow the function to continue).
|
||||
$pyvenvConfigPath = Join-Path -Resolve -Path $ConfigDir -ChildPath 'pyvenv.cfg' -ErrorAction Continue
|
||||
|
||||
# An empty map will be returned if no config file is found.
|
||||
$pyvenvConfig = @{ }
|
||||
|
||||
if ($pyvenvConfigPath) {
|
||||
|
||||
Write-Verbose "File exists, parse `key = value` lines"
|
||||
$pyvenvConfigContent = Get-Content -Path $pyvenvConfigPath
|
||||
|
||||
$pyvenvConfigContent | ForEach-Object {
|
||||
$keyval = $PSItem -split "\s*=\s*", 2
|
||||
if ($keyval[0] -and $keyval[1]) {
|
||||
$val = $keyval[1]
|
||||
|
||||
# Remove extraneous quotations around a string value.
|
||||
if ("'""".Contains($val.Substring(0, 1))) {
|
||||
$val = $val.Substring(1, $val.Length - 2)
|
||||
}
|
||||
|
||||
$pyvenvConfig[$keyval[0]] = $val
|
||||
Write-Verbose "Adding Key: '$($keyval[0])'='$val'"
|
||||
}
|
||||
}
|
||||
}
|
||||
return $pyvenvConfig
|
||||
}
|
||||
|
||||
|
||||
<# Begin Activate script --------------------------------------------------- #>
|
||||
|
||||
# Determine the containing directory of this script
|
||||
$VenvExecPath = Split-Path -Parent $MyInvocation.MyCommand.Definition
|
||||
$VenvExecDir = Get-Item -Path $VenvExecPath
|
||||
|
||||
Write-Verbose "Activation script is located in path: '$VenvExecPath'"
|
||||
Write-Verbose "VenvExecDir Fullname: '$($VenvExecDir.FullName)"
|
||||
Write-Verbose "VenvExecDir Name: '$($VenvExecDir.Name)"
|
||||
|
||||
# Set values required in priority: CmdLine, ConfigFile, Default
|
||||
# First, get the location of the virtual environment, it might not be
|
||||
# VenvExecDir if specified on the command line.
|
||||
if ($VenvDir) {
|
||||
Write-Verbose "VenvDir given as parameter, using '$VenvDir' to determine values"
|
||||
}
|
||||
else {
|
||||
Write-Verbose "VenvDir not given as a parameter, using parent directory name as VenvDir."
|
||||
$VenvDir = $VenvExecDir.Parent.FullName.TrimEnd("\\/")
|
||||
Write-Verbose "VenvDir=$VenvDir"
|
||||
}
|
||||
|
||||
# Next, read the `pyvenv.cfg` file to determine any required value such
|
||||
# as `prompt`.
|
||||
$pyvenvCfg = Get-PyVenvConfig -ConfigDir $VenvDir
|
||||
|
||||
# Next, set the prompt from the command line, or the config file, or
|
||||
# just use the name of the virtual environment folder.
|
||||
if ($Prompt) {
|
||||
Write-Verbose "Prompt specified as argument, using '$Prompt'"
|
||||
}
|
||||
else {
|
||||
Write-Verbose "Prompt not specified as argument to script, checking pyvenv.cfg value"
|
||||
if ($pyvenvCfg -and $pyvenvCfg['prompt']) {
|
||||
Write-Verbose " Setting based on value in pyvenv.cfg='$($pyvenvCfg['prompt'])'"
|
||||
$Prompt = $pyvenvCfg['prompt'];
|
||||
}
|
||||
else {
|
||||
Write-Verbose " Setting prompt based on parent's directory's name. (Is the directory name passed to venv module when creating the virtual environment)"
|
||||
Write-Verbose " Got leaf-name of $VenvDir='$(Split-Path -Path $venvDir -Leaf)'"
|
||||
$Prompt = Split-Path -Path $venvDir -Leaf
|
||||
}
|
||||
}
|
||||
|
||||
Write-Verbose "Prompt = '$Prompt'"
|
||||
Write-Verbose "VenvDir='$VenvDir'"
|
||||
|
||||
# Deactivate any currently active virtual environment, but leave the
|
||||
# deactivate function in place.
|
||||
deactivate -nondestructive
|
||||
|
||||
# Now set the environment variable VIRTUAL_ENV, used by many tools to determine
|
||||
# that there is an activated venv.
|
||||
$env:VIRTUAL_ENV = $VenvDir
|
||||
|
||||
if (-not $Env:VIRTUAL_ENV_DISABLE_PROMPT) {
|
||||
|
||||
Write-Verbose "Setting prompt to '$Prompt'"
|
||||
|
||||
# Set the prompt to include the env name
|
||||
# Make sure _OLD_VIRTUAL_PROMPT is global
|
||||
function global:_OLD_VIRTUAL_PROMPT { "" }
|
||||
Copy-Item -Path function:prompt -Destination function:_OLD_VIRTUAL_PROMPT
|
||||
New-Variable -Name _PYTHON_VENV_PROMPT_PREFIX -Description "Python virtual environment prompt prefix" -Scope Global -Option ReadOnly -Visibility Public -Value $Prompt
|
||||
|
||||
function global:prompt {
|
||||
Write-Host -NoNewline -ForegroundColor Green "($_PYTHON_VENV_PROMPT_PREFIX) "
|
||||
_OLD_VIRTUAL_PROMPT
|
||||
}
|
||||
$env:VIRTUAL_ENV_PROMPT = $Prompt
|
||||
}
|
||||
|
||||
# Clear PYTHONHOME
|
||||
if (Test-Path -Path Env:PYTHONHOME) {
|
||||
Copy-Item -Path Env:PYTHONHOME -Destination Env:_OLD_VIRTUAL_PYTHONHOME
|
||||
Remove-Item -Path Env:PYTHONHOME
|
||||
}
|
||||
|
||||
# Add the venv to the PATH
|
||||
Copy-Item -Path Env:PATH -Destination Env:_OLD_VIRTUAL_PATH
|
||||
$Env:PATH = "$VenvExecDir$([System.IO.Path]::PathSeparator)$Env:PATH"
|
||||
70
testclass/bin/activate
Normal file
70
testclass/bin/activate
Normal file
@ -0,0 +1,70 @@
|
||||
# This file must be used with "source bin/activate" *from bash*
|
||||
# You cannot run it directly
|
||||
|
||||
deactivate () {
|
||||
# reset old environment variables
|
||||
if [ -n "${_OLD_VIRTUAL_PATH:-}" ] ; then
|
||||
PATH="${_OLD_VIRTUAL_PATH:-}"
|
||||
export PATH
|
||||
unset _OLD_VIRTUAL_PATH
|
||||
fi
|
||||
if [ -n "${_OLD_VIRTUAL_PYTHONHOME:-}" ] ; then
|
||||
PYTHONHOME="${_OLD_VIRTUAL_PYTHONHOME:-}"
|
||||
export PYTHONHOME
|
||||
unset _OLD_VIRTUAL_PYTHONHOME
|
||||
fi
|
||||
|
||||
# Call hash to forget past commands. Without forgetting
|
||||
# past commands the $PATH changes we made may not be respected
|
||||
hash -r 2> /dev/null
|
||||
|
||||
if [ -n "${_OLD_VIRTUAL_PS1:-}" ] ; then
|
||||
PS1="${_OLD_VIRTUAL_PS1:-}"
|
||||
export PS1
|
||||
unset _OLD_VIRTUAL_PS1
|
||||
fi
|
||||
|
||||
unset VIRTUAL_ENV
|
||||
unset VIRTUAL_ENV_PROMPT
|
||||
if [ ! "${1:-}" = "nondestructive" ] ; then
|
||||
# Self destruct!
|
||||
unset -f deactivate
|
||||
fi
|
||||
}
|
||||
|
||||
# unset irrelevant variables
|
||||
deactivate nondestructive
|
||||
|
||||
# on Windows, a path can contain colons and backslashes and has to be converted:
|
||||
if [ "${OSTYPE:-}" = "cygwin" ] || [ "${OSTYPE:-}" = "msys" ] ; then
|
||||
# transform D:\path\to\venv to /d/path/to/venv on MSYS
|
||||
# and to /cygdrive/d/path/to/venv on Cygwin
|
||||
export VIRTUAL_ENV=$(cygpath /home/fgras-ca/AIagent2/testclass)
|
||||
else
|
||||
# use the path as-is
|
||||
export VIRTUAL_ENV=/home/fgras-ca/AIagent2/testclass
|
||||
fi
|
||||
|
||||
_OLD_VIRTUAL_PATH="$PATH"
|
||||
PATH="$VIRTUAL_ENV/"bin":$PATH"
|
||||
export PATH
|
||||
|
||||
# unset PYTHONHOME if set
|
||||
# this will fail if PYTHONHOME is set to the empty string (which is bad anyway)
|
||||
# could use `if (set -u; : $PYTHONHOME) ;` in bash
|
||||
if [ -n "${PYTHONHOME:-}" ] ; then
|
||||
_OLD_VIRTUAL_PYTHONHOME="${PYTHONHOME:-}"
|
||||
unset PYTHONHOME
|
||||
fi
|
||||
|
||||
if [ -z "${VIRTUAL_ENV_DISABLE_PROMPT:-}" ] ; then
|
||||
_OLD_VIRTUAL_PS1="${PS1:-}"
|
||||
PS1='(testclass) '"${PS1:-}"
|
||||
export PS1
|
||||
VIRTUAL_ENV_PROMPT='(testclass) '
|
||||
export VIRTUAL_ENV_PROMPT
|
||||
fi
|
||||
|
||||
# Call hash to forget past commands. Without forgetting
|
||||
# past commands the $PATH changes we made may not be respected
|
||||
hash -r 2> /dev/null
|
||||
27
testclass/bin/activate.csh
Normal file
27
testclass/bin/activate.csh
Normal file
@ -0,0 +1,27 @@
|
||||
# This file must be used with "source bin/activate.csh" *from csh*.
|
||||
# You cannot run it directly.
|
||||
|
||||
# Created by Davide Di Blasi <davidedb@gmail.com>.
|
||||
# Ported to Python 3.3 venv by Andrew Svetlov <andrew.svetlov@gmail.com>
|
||||
|
||||
alias deactivate 'test $?_OLD_VIRTUAL_PATH != 0 && setenv PATH "$_OLD_VIRTUAL_PATH" && unset _OLD_VIRTUAL_PATH; rehash; test $?_OLD_VIRTUAL_PROMPT != 0 && set prompt="$_OLD_VIRTUAL_PROMPT" && unset _OLD_VIRTUAL_PROMPT; unsetenv VIRTUAL_ENV; unsetenv VIRTUAL_ENV_PROMPT; test "\!:*" != "nondestructive" && unalias deactivate'
|
||||
|
||||
# Unset irrelevant variables.
|
||||
deactivate nondestructive
|
||||
|
||||
setenv VIRTUAL_ENV /home/fgras-ca/AIagent2/testclass
|
||||
|
||||
set _OLD_VIRTUAL_PATH="$PATH"
|
||||
setenv PATH "$VIRTUAL_ENV/"bin":$PATH"
|
||||
|
||||
|
||||
set _OLD_VIRTUAL_PROMPT="$prompt"
|
||||
|
||||
if (! "$?VIRTUAL_ENV_DISABLE_PROMPT") then
|
||||
set prompt = '(testclass) '"$prompt"
|
||||
setenv VIRTUAL_ENV_PROMPT '(testclass) '
|
||||
endif
|
||||
|
||||
alias pydoc python -m pydoc
|
||||
|
||||
rehash
|
||||
69
testclass/bin/activate.fish
Normal file
69
testclass/bin/activate.fish
Normal file
@ -0,0 +1,69 @@
|
||||
# This file must be used with "source <venv>/bin/activate.fish" *from fish*
|
||||
# (https://fishshell.com/). You cannot run it directly.
|
||||
|
||||
function deactivate -d "Exit virtual environment and return to normal shell environment"
|
||||
# reset old environment variables
|
||||
if test -n "$_OLD_VIRTUAL_PATH"
|
||||
set -gx PATH $_OLD_VIRTUAL_PATH
|
||||
set -e _OLD_VIRTUAL_PATH
|
||||
end
|
||||
if test -n "$_OLD_VIRTUAL_PYTHONHOME"
|
||||
set -gx PYTHONHOME $_OLD_VIRTUAL_PYTHONHOME
|
||||
set -e _OLD_VIRTUAL_PYTHONHOME
|
||||
end
|
||||
|
||||
if test -n "$_OLD_FISH_PROMPT_OVERRIDE"
|
||||
set -e _OLD_FISH_PROMPT_OVERRIDE
|
||||
# prevents error when using nested fish instances (Issue #93858)
|
||||
if functions -q _old_fish_prompt
|
||||
functions -e fish_prompt
|
||||
functions -c _old_fish_prompt fish_prompt
|
||||
functions -e _old_fish_prompt
|
||||
end
|
||||
end
|
||||
|
||||
set -e VIRTUAL_ENV
|
||||
set -e VIRTUAL_ENV_PROMPT
|
||||
if test "$argv[1]" != "nondestructive"
|
||||
# Self-destruct!
|
||||
functions -e deactivate
|
||||
end
|
||||
end
|
||||
|
||||
# Unset irrelevant variables.
|
||||
deactivate nondestructive
|
||||
|
||||
set -gx VIRTUAL_ENV /home/fgras-ca/AIagent2/testclass
|
||||
|
||||
set -gx _OLD_VIRTUAL_PATH $PATH
|
||||
set -gx PATH "$VIRTUAL_ENV/"bin $PATH
|
||||
|
||||
# Unset PYTHONHOME if set.
|
||||
if set -q PYTHONHOME
|
||||
set -gx _OLD_VIRTUAL_PYTHONHOME $PYTHONHOME
|
||||
set -e PYTHONHOME
|
||||
end
|
||||
|
||||
if test -z "$VIRTUAL_ENV_DISABLE_PROMPT"
|
||||
# fish uses a function instead of an env var to generate the prompt.
|
||||
|
||||
# Save the current fish_prompt function as the function _old_fish_prompt.
|
||||
functions -c fish_prompt _old_fish_prompt
|
||||
|
||||
# With the original prompt function renamed, we can override with our own.
|
||||
function fish_prompt
|
||||
# Save the return status of the last command.
|
||||
set -l old_status $status
|
||||
|
||||
# Output the venv prompt; color taken from the blue of the Python logo.
|
||||
printf "%s%s%s" (set_color 4B8BBE) '(testclass) ' (set_color normal)
|
||||
|
||||
# Restore the return status of the previous command.
|
||||
echo "exit $old_status" | .
|
||||
# Output the original/"old" prompt.
|
||||
_old_fish_prompt
|
||||
end
|
||||
|
||||
set -gx _OLD_FISH_PROMPT_OVERRIDE "$VIRTUAL_ENV"
|
||||
set -gx VIRTUAL_ENV_PROMPT '(testclass) '
|
||||
end
|
||||
8
testclass/bin/normalizer
Executable file
8
testclass/bin/normalizer
Executable file
@ -0,0 +1,8 @@
|
||||
#!/home/fgras-ca/AIagent2/testclass/bin/python3
|
||||
# -*- coding: utf-8 -*-
|
||||
import re
|
||||
import sys
|
||||
from charset_normalizer import cli
|
||||
if __name__ == '__main__':
|
||||
sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
|
||||
sys.exit(cli.cli_detect())
|
||||
8
testclass/bin/pip
Executable file
8
testclass/bin/pip
Executable file
@ -0,0 +1,8 @@
|
||||
#!/home/fgras-ca/AIagent2/testclass/bin/python3
|
||||
# -*- coding: utf-8 -*-
|
||||
import re
|
||||
import sys
|
||||
from pip._internal.cli.main import main
|
||||
if __name__ == '__main__':
|
||||
sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
|
||||
sys.exit(main())
|
||||
8
testclass/bin/pip3
Executable file
8
testclass/bin/pip3
Executable file
@ -0,0 +1,8 @@
|
||||
#!/home/fgras-ca/AIagent2/testclass/bin/python3
|
||||
# -*- coding: utf-8 -*-
|
||||
import re
|
||||
import sys
|
||||
from pip._internal.cli.main import main
|
||||
if __name__ == '__main__':
|
||||
sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
|
||||
sys.exit(main())
|
||||
8
testclass/bin/pip3.12
Executable file
8
testclass/bin/pip3.12
Executable file
@ -0,0 +1,8 @@
|
||||
#!/home/fgras-ca/AIagent2/testclass/bin/python3
|
||||
# -*- coding: utf-8 -*-
|
||||
import re
|
||||
import sys
|
||||
from pip._internal.cli.main import main
|
||||
if __name__ == '__main__':
|
||||
sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
|
||||
sys.exit(main())
|
||||
1
testclass/bin/python
Symbolic link
1
testclass/bin/python
Symbolic link
@ -0,0 +1 @@
|
||||
python3
|
||||
1
testclass/bin/python3
Symbolic link
1
testclass/bin/python3
Symbolic link
@ -0,0 +1 @@
|
||||
/usr/bin/python3
|
||||
1
testclass/bin/python3.12
Symbolic link
1
testclass/bin/python3.12
Symbolic link
@ -0,0 +1 @@
|
||||
python3
|
||||
@ -0,0 +1 @@
|
||||
pip
|
||||
@ -0,0 +1,44 @@
|
||||
Zope Public License (ZPL) Version 2.1
|
||||
|
||||
A copyright notice accompanies this license document that identifies the
|
||||
copyright holders.
|
||||
|
||||
This license has been certified as open source. It has also been designated as
|
||||
GPL compatible by the Free Software Foundation (FSF).
|
||||
|
||||
Redistribution and use in source and binary forms, with or without
|
||||
modification, are permitted provided that the following conditions are met:
|
||||
|
||||
1. Redistributions in source code must retain the accompanying copyright
|
||||
notice, this list of conditions, and the following disclaimer.
|
||||
|
||||
2. Redistributions in binary form must reproduce the accompanying copyright
|
||||
notice, this list of conditions, and the following disclaimer in the
|
||||
documentation and/or other materials provided with the distribution.
|
||||
|
||||
3. Names of the copyright holders must not be used to endorse or promote
|
||||
products derived from this software without prior written permission from the
|
||||
copyright holders.
|
||||
|
||||
4. The right to distribute this software or to use it for any purpose does not
|
||||
give you the right to use Servicemarks (sm) or Trademarks (tm) of the
|
||||
copyright
|
||||
holders. Use of them is covered by separate agreement with the copyright
|
||||
holders.
|
||||
|
||||
5. If any files are modified, you must cause the modified files to carry
|
||||
prominent notices stating that you changed the files and the date of any
|
||||
change.
|
||||
|
||||
Disclaimer
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ``AS IS'' AND ANY EXPRESSED
|
||||
OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
||||
OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO
|
||||
EVENT SHALL THE COPYRIGHT HOLDERS BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||||
INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
|
||||
LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
|
||||
NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
|
||||
EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,22 @@
|
||||
DateTime-5.5.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
|
||||
DateTime-5.5.dist-info/LICENSE.txt,sha256=PmcdsR32h1FswdtbPWXkqjg-rKPCDOo_r1Og9zNdCjw,2070
|
||||
DateTime-5.5.dist-info/METADATA,sha256=W1k0PqPJ6SU6QTJAu40JPtHK8XeQRL0GGEpfVGPjWGI,33735
|
||||
DateTime-5.5.dist-info/RECORD,,
|
||||
DateTime-5.5.dist-info/REQUESTED,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
||||
DateTime-5.5.dist-info/WHEEL,sha256=oiQVh_5PnQM0E3gPdiz09WCNmwiHDMaGer_elqB3coM,92
|
||||
DateTime-5.5.dist-info/top_level.txt,sha256=iVdUvuV_RIkkMzsnPGNfwojRWvuonInryaK3hA5Hh0o,9
|
||||
DateTime/DateTime.py,sha256=65LbTcnrCSsDPGoGLVkk7NC3H8Kq-PjkC1fQVR33gE8,71364
|
||||
DateTime/DateTime.txt,sha256=KZFzxoQItLsar1ZDd2vZN74Y6L4a04H8jXMwqc8KjmY,22487
|
||||
DateTime/__init__.py,sha256=trlFzEmNkmUpxZT7krPSVDayDK1bRxToccg3CcCF8wg,714
|
||||
DateTime/__pycache__/DateTime.cpython-312.pyc,,
|
||||
DateTime/__pycache__/__init__.cpython-312.pyc,,
|
||||
DateTime/__pycache__/interfaces.cpython-312.pyc,,
|
||||
DateTime/__pycache__/pytz_support.cpython-312.pyc,,
|
||||
DateTime/interfaces.py,sha256=n47sexf1eQ6YMdYB_60PgHtSzYIj4FND-RmHFiNpm1E,12187
|
||||
DateTime/pytz.txt,sha256=9Phns9ESXs9MaOKxXztX6sJ09QczGxsbYoSRSllKUfk,5619
|
||||
DateTime/pytz_support.py,sha256=inR1SO0X17fp9C2GsRw99S_MhxKiEt5dOV3-TGsBxDI,11853
|
||||
DateTime/tests/__init__.py,sha256=H7Ixo1xp-8BlJ65u14hk5i_TKEmETyi2FmLMD6H-mpo,683
|
||||
DateTime/tests/__pycache__/__init__.cpython-312.pyc,,
|
||||
DateTime/tests/__pycache__/test_datetime.cpython-312.pyc,,
|
||||
DateTime/tests/julian_testdata.txt,sha256=qxvLvabVB9ayhh5UHBvPhuqW5mRL_lizzbUh6lc3d4I,1397
|
||||
DateTime/tests/test_datetime.py,sha256=dsrfAqQpz1od1bOVPvSYfZAlduJpJIpc2F_hdN7WRAU,30385
|
||||
@ -0,0 +1,5 @@
|
||||
Wheel-Version: 1.0
|
||||
Generator: bdist_wheel (0.42.0)
|
||||
Root-Is-Purelib: true
|
||||
Tag: py3-none-any
|
||||
|
||||
@ -0,0 +1 @@
|
||||
DateTime
|
||||
1948
testclass/lib/python3.12/site-packages/DateTime/DateTime.py
Normal file
1948
testclass/lib/python3.12/site-packages/DateTime/DateTime.py
Normal file
File diff suppressed because it is too large
Load Diff
785
testclass/lib/python3.12/site-packages/DateTime/DateTime.txt
Normal file
785
testclass/lib/python3.12/site-packages/DateTime/DateTime.txt
Normal file
@ -0,0 +1,785 @@
|
||||
The DateTime package
|
||||
====================
|
||||
|
||||
Encapsulation of date/time values.
|
||||
|
||||
|
||||
Function Timezones()
|
||||
--------------------
|
||||
|
||||
Returns the list of recognized timezone names:
|
||||
|
||||
>>> from DateTime import Timezones
|
||||
>>> zones = set(Timezones())
|
||||
|
||||
Almost all of the standard pytz timezones are included, with the exception
|
||||
of some commonly-used but ambiguous abbreviations, where historical Zope
|
||||
usage conflicts with the name used by pytz:
|
||||
|
||||
>>> import pytz
|
||||
>>> [x for x in pytz.all_timezones if x not in zones]
|
||||
['CET', 'EET', 'EST', 'MET', 'MST', 'WET']
|
||||
|
||||
Class DateTime
|
||||
--------------
|
||||
|
||||
DateTime objects represent instants in time and provide interfaces for
|
||||
controlling its representation without affecting the absolute value of
|
||||
the object.
|
||||
|
||||
DateTime objects may be created from a wide variety of string or
|
||||
numeric data, or may be computed from other DateTime objects.
|
||||
DateTimes support the ability to convert their representations to many
|
||||
major timezones, as well as the ability to create a DateTime object
|
||||
in the context of a given timezone.
|
||||
|
||||
DateTime objects provide partial numerical behavior:
|
||||
|
||||
* Two date-time objects can be subtracted to obtain a time, in days
|
||||
between the two.
|
||||
|
||||
* A date-time object and a positive or negative number may be added to
|
||||
obtain a new date-time object that is the given number of days later
|
||||
than the input date-time object.
|
||||
|
||||
* A positive or negative number and a date-time object may be added to
|
||||
obtain a new date-time object that is the given number of days later
|
||||
than the input date-time object.
|
||||
|
||||
* A positive or negative number may be subtracted from a date-time
|
||||
object to obtain a new date-time object that is the given number of
|
||||
days earlier than the input date-time object.
|
||||
|
||||
DateTime objects may be converted to integer, long, or float numbers
|
||||
of days since January 1, 1901, using the standard int, long, and float
|
||||
functions (Compatibility Note: int, long and float return the number
|
||||
of days since 1901 in GMT rather than local machine timezone).
|
||||
DateTime objects also provide access to their value in a float format
|
||||
usable with the Python time module, provided that the value of the
|
||||
object falls in the range of the epoch-based time module.
|
||||
|
||||
A DateTime object should be considered immutable; all conversion and numeric
|
||||
operations return a new DateTime object rather than modify the current object.
|
||||
|
||||
A DateTime object always maintains its value as an absolute UTC time,
|
||||
and is represented in the context of some timezone based on the
|
||||
arguments used to create the object. A DateTime object's methods
|
||||
return values based on the timezone context.
|
||||
|
||||
Note that in all cases the local machine timezone is used for
|
||||
representation if no timezone is specified.
|
||||
|
||||
Constructor for DateTime
|
||||
------------------------
|
||||
|
||||
DateTime() returns a new date-time object. DateTimes may be created
|
||||
with from zero to seven arguments:
|
||||
|
||||
* If the function is called with no arguments, then the current date/
|
||||
time is returned, represented in the timezone of the local machine.
|
||||
|
||||
* If the function is invoked with a single string argument which is a
|
||||
recognized timezone name, an object representing the current time is
|
||||
returned, represented in the specified timezone.
|
||||
|
||||
* If the function is invoked with a single string argument
|
||||
representing a valid date/time, an object representing that date/
|
||||
time will be returned.
|
||||
|
||||
As a general rule, any date-time representation that is recognized
|
||||
and unambiguous to a resident of North America is acceptable. (The
|
||||
reason for this qualification is that in North America, a date like:
|
||||
2/1/1994 is interpreted as February 1, 1994, while in some parts of
|
||||
the world, it is interpreted as January 2, 1994.) A date/ time
|
||||
string consists of two components, a date component and an optional
|
||||
time component, separated by one or more spaces. If the time
|
||||
component is omitted, 12:00am is assumed.
|
||||
|
||||
Any recognized timezone name specified as the final element of the
|
||||
date/time string will be used for computing the date/time value.
|
||||
(If you create a DateTime with the string,
|
||||
"Mar 9, 1997 1:45pm US/Pacific", the value will essentially be the
|
||||
same as if you had captured time.time() at the specified date and
|
||||
time on a machine in that timezone). If no timezone is passed, then
|
||||
the timezone configured on the local machine will be used, **except**
|
||||
that if the date format matches ISO 8601 ('YYYY-MM-DD'), the instance
|
||||
will use UTC / GMT+0 as the timezone.
|
||||
|
||||
o Returns current date/time, represented in US/Eastern:
|
||||
|
||||
>>> from DateTime import DateTime
|
||||
>>> e = DateTime('US/Eastern')
|
||||
>>> e.timezone()
|
||||
'US/Eastern'
|
||||
|
||||
o Returns specified time, represented in local machine zone:
|
||||
|
||||
>>> x = DateTime('1997/3/9 1:45pm')
|
||||
>>> x.parts() # doctest: +ELLIPSIS
|
||||
(1997, 3, 9, 13, 45, ...)
|
||||
|
||||
o Specified time in local machine zone, verbose format:
|
||||
|
||||
>>> y = DateTime('Mar 9, 1997 13:45:00')
|
||||
>>> y.parts() # doctest: +ELLIPSIS
|
||||
(1997, 3, 9, 13, 45, ...)
|
||||
>>> y == x
|
||||
True
|
||||
|
||||
o Specified time in UTC via ISO 8601 rule:
|
||||
|
||||
>>> z = DateTime('2014-03-24')
|
||||
>>> z.parts() # doctest: +ELLIPSIS
|
||||
(2014, 3, 24, 0, 0, ...)
|
||||
>>> z.timezone()
|
||||
'GMT+0'
|
||||
|
||||
The date component consists of year, month, and day values. The
|
||||
year value must be a one-, two-, or four-digit integer. If a one-
|
||||
or two-digit year is used, the year is assumed to be in the
|
||||
twentieth century. The month may an integer, from 1 to 12, a month
|
||||
name, or a month abbreviation, where a period may optionally follow
|
||||
the abbreviation. The day must be an integer from 1 to the number of
|
||||
days in the month. The year, month, and day values may be separated
|
||||
by periods, hyphens, forward slashes, or spaces. Extra spaces are
|
||||
permitted around the delimiters. Year, month, and day values may be
|
||||
given in any order as long as it is possible to distinguish the
|
||||
components. If all three components are numbers that are less than
|
||||
13, then a month-day-year ordering is assumed.
|
||||
|
||||
The time component consists of hour, minute, and second values
|
||||
separated by colons. The hour value must be an integer between 0
|
||||
and 23 inclusively. The minute value must be an integer between 0
|
||||
and 59 inclusively. The second value may be an integer value
|
||||
between 0 and 59.999 inclusively. The second value or both the
|
||||
minute and second values may be omitted. The time may be followed
|
||||
by am or pm in upper or lower case, in which case a 12-hour clock is
|
||||
assumed.
|
||||
|
||||
* If the DateTime function is invoked with a single numeric argument,
|
||||
the number is assumed to be either a floating point value such as
|
||||
that returned by time.time(), or a number of days after January 1,
|
||||
1901 00:00:00 UTC.
|
||||
|
||||
A DateTime object is returned that represents either the GMT value
|
||||
of the time.time() float represented in the local machine's
|
||||
timezone, or that number of days after January 1, 1901. Note that
|
||||
the number of days after 1901 need to be expressed from the
|
||||
viewpoint of the local machine's timezone. A negative argument will
|
||||
yield a date-time value before 1901.
|
||||
|
||||
* If the function is invoked with two numeric arguments, then the
|
||||
first is taken to be an integer year and the second argument is
|
||||
taken to be an offset in days from the beginning of the year, in the
|
||||
context of the local machine timezone. The date-time value returned
|
||||
is the given offset number of days from the beginning of the given
|
||||
year, represented in the timezone of the local machine. The offset
|
||||
may be positive or negative. Two-digit years are assumed to be in
|
||||
the twentieth century.
|
||||
|
||||
* If the function is invoked with two arguments, the first a float
|
||||
representing a number of seconds past the epoch in GMT (such as
|
||||
those returned by time.time()) and the second a string naming a
|
||||
recognized timezone, a DateTime with a value of that GMT time will
|
||||
be returned, represented in the given timezone.
|
||||
|
||||
>>> import time
|
||||
>>> t = time.time()
|
||||
|
||||
Time t represented as US/Eastern:
|
||||
|
||||
>>> now_east = DateTime(t, 'US/Eastern')
|
||||
|
||||
Time t represented as US/Pacific:
|
||||
|
||||
>>> now_west = DateTime(t, 'US/Pacific')
|
||||
|
||||
Only their representations are different:
|
||||
|
||||
>>> now_east.equalTo(now_west)
|
||||
True
|
||||
|
||||
* If the function is invoked with three or more numeric arguments,
|
||||
then the first is taken to be an integer year, the second is taken
|
||||
to be an integer month, and the third is taken to be an integer day.
|
||||
If the combination of values is not valid, then a DateTimeError is
|
||||
raised. One- or two-digit years up to 69 are assumed to be in the
|
||||
21st century, whereas values 70-99 are assumed to be 20th century.
|
||||
The fourth, fifth, and sixth arguments are floating point, positive
|
||||
or negative offsets in units of hours, minutes, and days, and
|
||||
default to zero if not given. An optional string may be given as
|
||||
the final argument to indicate timezone (the effect of this is as if
|
||||
you had taken the value of time.time() at that time on a machine in
|
||||
the specified timezone).
|
||||
|
||||
If a string argument passed to the DateTime constructor cannot be
|
||||
parsed, it will raise SyntaxError. Invalid date, time, or
|
||||
timezone components will raise a DateTimeError.
|
||||
|
||||
The module function Timezones() will return a list of the timezones
|
||||
recognized by the DateTime module. Recognition of timezone names is
|
||||
case-insensitive.
|
||||
|
||||
Instance Methods for DateTime (IDateTime interface)
|
||||
---------------------------------------------------
|
||||
|
||||
Conversion and comparison methods
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
* ``timeTime()`` returns the date/time as a floating-point number in
|
||||
UTC, in the format used by the Python time module. Note that it is
|
||||
possible to create date /time values with DateTime that have no
|
||||
meaningful value to the time module, and in such cases a
|
||||
DateTimeError is raised. A DateTime object's value must generally
|
||||
be between Jan 1, 1970 (or your local machine epoch) and Jan 2038 to
|
||||
produce a valid time.time() style value.
|
||||
|
||||
>>> dt = DateTime('Mar 9, 1997 13:45:00 US/Eastern')
|
||||
>>> dt.timeTime()
|
||||
857933100.0
|
||||
|
||||
>>> DateTime('2040/01/01 UTC').timeTime()
|
||||
2208988800.0
|
||||
|
||||
>>> DateTime('1900/01/01 UTC').timeTime()
|
||||
-2208988800.0
|
||||
|
||||
* ``toZone(z)`` returns a DateTime with the value as the current
|
||||
object, represented in the indicated timezone:
|
||||
|
||||
>>> dt.toZone('UTC')
|
||||
DateTime('1997/03/09 18:45:00 UTC')
|
||||
|
||||
>>> dt.toZone('UTC').equalTo(dt)
|
||||
True
|
||||
|
||||
* ``isFuture()`` returns true if this object represents a date/time
|
||||
later than the time of the call:
|
||||
|
||||
>>> dt.isFuture()
|
||||
False
|
||||
>>> DateTime('Jan 1 3000').isFuture() # not time-machine safe!
|
||||
True
|
||||
|
||||
* ``isPast()`` returns true if this object represents a date/time
|
||||
earlier than the time of the call:
|
||||
|
||||
>>> dt.isPast()
|
||||
True
|
||||
>>> DateTime('Jan 1 3000').isPast() # not time-machine safe!
|
||||
False
|
||||
|
||||
* ``isCurrentYear()`` returns true if this object represents a
|
||||
date/time that falls within the current year, in the context of this
|
||||
object's timezone representation:
|
||||
|
||||
>>> dt.isCurrentYear()
|
||||
False
|
||||
>>> DateTime().isCurrentYear()
|
||||
True
|
||||
|
||||
* ``isCurrentMonth()`` returns true if this object represents a
|
||||
date/time that falls within the current month, in the context of
|
||||
this object's timezone representation:
|
||||
|
||||
>>> dt.isCurrentMonth()
|
||||
False
|
||||
>>> DateTime().isCurrentMonth()
|
||||
True
|
||||
|
||||
* ``isCurrentDay()`` returns true if this object represents a
|
||||
date/time that falls within the current day, in the context of this
|
||||
object's timezone representation:
|
||||
|
||||
>>> dt.isCurrentDay()
|
||||
False
|
||||
>>> DateTime().isCurrentDay()
|
||||
True
|
||||
|
||||
* ``isCurrentHour()`` returns true if this object represents a
|
||||
date/time that falls within the current hour, in the context of this
|
||||
object's timezone representation:
|
||||
|
||||
>>> dt.isCurrentHour()
|
||||
False
|
||||
|
||||
>>> DateTime().isCurrentHour()
|
||||
True
|
||||
|
||||
* ``isCurrentMinute()`` returns true if this object represents a
|
||||
date/time that falls within the current minute, in the context of
|
||||
this object's timezone representation:
|
||||
|
||||
>>> dt.isCurrentMinute()
|
||||
False
|
||||
>>> DateTime().isCurrentMinute()
|
||||
True
|
||||
|
||||
* ``isLeapYear()`` returns true if the current year (in the context of
|
||||
the object's timezone) is a leap year:
|
||||
|
||||
>>> dt.isLeapYear()
|
||||
False
|
||||
>>> DateTime('Mar 8 2004').isLeapYear()
|
||||
True
|
||||
|
||||
* ``earliestTime()`` returns a new DateTime object that represents the
|
||||
earliest possible time (in whole seconds) that still falls within
|
||||
the current object's day, in the object's timezone context:
|
||||
|
||||
>>> dt.earliestTime()
|
||||
DateTime('1997/03/09 00:00:00 US/Eastern')
|
||||
|
||||
* ``latestTime()`` return a new DateTime object that represents the
|
||||
latest possible time (in whole seconds) that still falls within the
|
||||
current object's day, in the object's timezone context
|
||||
|
||||
>>> dt.latestTime()
|
||||
DateTime('1997/03/09 23:59:59 US/Eastern')
|
||||
|
||||
Component access
|
||||
~~~~~~~~~~~~~~~~
|
||||
|
||||
* ``parts()`` returns a tuple containing the calendar year, month,
|
||||
day, hour, minute second and timezone of the object
|
||||
|
||||
>>> dt.parts() # doctest: +ELLIPSIS
|
||||
(1997, 3, 9, 13, 45, ... 'US/Eastern')
|
||||
|
||||
* ``timezone()`` returns the timezone in which the object is represented:
|
||||
|
||||
>>> dt.timezone() in Timezones()
|
||||
True
|
||||
|
||||
* ``tzoffset()`` returns the timezone offset for the objects timezone:
|
||||
|
||||
>>> dt.tzoffset()
|
||||
-18000
|
||||
|
||||
* ``year()`` returns the calendar year of the object:
|
||||
|
||||
>>> dt.year()
|
||||
1997
|
||||
|
||||
* ``month()`` returns the month of the object as an integer:
|
||||
|
||||
>>> dt.month()
|
||||
3
|
||||
|
||||
* ``Month()`` returns the full month name:
|
||||
|
||||
>>> dt.Month()
|
||||
'March'
|
||||
|
||||
* ``aMonth()`` returns the abbreviated month name:
|
||||
|
||||
>>> dt.aMonth()
|
||||
'Mar'
|
||||
|
||||
* ``pMonth()`` returns the abbreviated (with period) month name:
|
||||
|
||||
>>> dt.pMonth()
|
||||
'Mar.'
|
||||
|
||||
* ``day()`` returns the integer day:
|
||||
|
||||
>>> dt.day()
|
||||
9
|
||||
|
||||
* ``Day()`` returns the full name of the day of the week:
|
||||
|
||||
>>> dt.Day()
|
||||
'Sunday'
|
||||
|
||||
* ``dayOfYear()`` returns the day of the year, in context of the
|
||||
timezone representation of the object:
|
||||
|
||||
>>> dt.dayOfYear()
|
||||
68
|
||||
|
||||
* ``aDay()`` returns the abbreviated name of the day of the week:
|
||||
|
||||
>>> dt.aDay()
|
||||
'Sun'
|
||||
|
||||
* ``pDay()`` returns the abbreviated (with period) name of the day of
|
||||
the week:
|
||||
|
||||
>>> dt.pDay()
|
||||
'Sun.'
|
||||
|
||||
* ``dow()`` returns the integer day of the week, where Sunday is 0:
|
||||
|
||||
>>> dt.dow()
|
||||
0
|
||||
|
||||
* ``dow_1()`` returns the integer day of the week, where sunday is 1:
|
||||
|
||||
>>> dt.dow_1()
|
||||
1
|
||||
|
||||
* ``h_12()`` returns the 12-hour clock representation of the hour:
|
||||
|
||||
>>> dt.h_12()
|
||||
1
|
||||
|
||||
* ``h_24()`` returns the 24-hour clock representation of the hour:
|
||||
|
||||
>>> dt.h_24()
|
||||
13
|
||||
|
||||
* ``ampm()`` returns the appropriate time modifier (am or pm):
|
||||
|
||||
>>> dt.ampm()
|
||||
'pm'
|
||||
|
||||
* ``hour()`` returns the 24-hour clock representation of the hour:
|
||||
|
||||
>>> dt.hour()
|
||||
13
|
||||
|
||||
* ``minute()`` returns the minute:
|
||||
|
||||
>>> dt.minute()
|
||||
45
|
||||
|
||||
* ``second()`` returns the second:
|
||||
|
||||
>>> dt.second() == 0
|
||||
True
|
||||
|
||||
* ``millis()`` returns the milliseconds since the epoch in GMT.
|
||||
|
||||
>>> dt.millis() == 857933100000
|
||||
True
|
||||
|
||||
strftime()
|
||||
~~~~~~~~~~
|
||||
|
||||
See ``tests/test_datetime.py``.
|
||||
|
||||
General formats from previous DateTime
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
* ``Date()`` return the date string for the object:
|
||||
|
||||
>>> dt.Date()
|
||||
'1997/03/09'
|
||||
|
||||
* ``Time()`` returns the time string for an object to the nearest
|
||||
second:
|
||||
|
||||
>>> dt.Time()
|
||||
'13:45:00'
|
||||
|
||||
* ``TimeMinutes()`` returns the time string for an object not showing
|
||||
seconds:
|
||||
|
||||
>>> dt.TimeMinutes()
|
||||
'13:45'
|
||||
|
||||
* ``AMPM()`` returns the time string for an object to the nearest second:
|
||||
|
||||
>>> dt.AMPM()
|
||||
'01:45:00 pm'
|
||||
|
||||
* ``AMPMMinutes()`` returns the time string for an object not showing
|
||||
seconds:
|
||||
|
||||
>>> dt.AMPMMinutes()
|
||||
'01:45 pm'
|
||||
|
||||
* ``PreciseTime()`` returns the time string for the object:
|
||||
|
||||
>>> dt.PreciseTime()
|
||||
'13:45:00.000'
|
||||
|
||||
* ``PreciseAMPM()`` returns the time string for the object:
|
||||
|
||||
>>> dt.PreciseAMPM()
|
||||
'01:45:00.000 pm'
|
||||
|
||||
* ``yy()`` returns the calendar year as a 2 digit string
|
||||
|
||||
>>> dt.yy()
|
||||
'97'
|
||||
|
||||
* ``mm()`` returns the month as a 2 digit string
|
||||
|
||||
>>> dt.mm()
|
||||
'03'
|
||||
|
||||
* ``dd()`` returns the day as a 2 digit string:
|
||||
|
||||
>>> dt.dd()
|
||||
'09'
|
||||
|
||||
* ``rfc822()`` returns the date in RFC 822 format:
|
||||
|
||||
>>> dt.rfc822()
|
||||
'Sun, 09 Mar 1997 13:45:00 -0500'
|
||||
|
||||
New formats
|
||||
~~~~~~~~~~~
|
||||
|
||||
* ``fCommon()`` returns a string representing the object's value in
|
||||
the format: March 9, 1997 1:45 pm:
|
||||
|
||||
>>> dt.fCommon()
|
||||
'March 9, 1997 1:45 pm'
|
||||
|
||||
* ``fCommonZ()`` returns a string representing the object's value in
|
||||
the format: March 9, 1997 1:45 pm US/Eastern:
|
||||
|
||||
>>> dt.fCommonZ()
|
||||
'March 9, 1997 1:45 pm US/Eastern'
|
||||
|
||||
* ``aCommon()`` returns a string representing the object's value in
|
||||
the format: Mar 9, 1997 1:45 pm:
|
||||
|
||||
>>> dt.aCommon()
|
||||
'Mar 9, 1997 1:45 pm'
|
||||
|
||||
* ``aCommonZ()`` return a string representing the object's value in
|
||||
the format: Mar 9, 1997 1:45 pm US/Eastern:
|
||||
|
||||
>>> dt.aCommonZ()
|
||||
'Mar 9, 1997 1:45 pm US/Eastern'
|
||||
|
||||
* ``pCommon()`` returns a string representing the object's value in
|
||||
the format Mar. 9, 1997 1:45 pm:
|
||||
|
||||
>>> dt.pCommon()
|
||||
'Mar. 9, 1997 1:45 pm'
|
||||
|
||||
* ``pCommonZ()`` returns a string representing the object's value in
|
||||
the format: Mar. 9, 1997 1:45 pm US/Eastern:
|
||||
|
||||
>>> dt.pCommonZ()
|
||||
'Mar. 9, 1997 1:45 pm US/Eastern'
|
||||
|
||||
* ``ISO()`` returns a string with the date/time in ISO format. Note:
|
||||
this is not ISO 8601-format! See the ISO8601 and HTML4 methods below
|
||||
for ISO 8601-compliant output. Dates are output as: YYYY-MM-DD HH:MM:SS
|
||||
|
||||
>>> dt.ISO()
|
||||
'1997-03-09 13:45:00'
|
||||
|
||||
* ``ISO8601()`` returns the object in ISO 8601-compatible format
|
||||
containing the date, time with seconds-precision and the time zone
|
||||
identifier - see http://www.w3.org/TR/NOTE-datetime. Dates are
|
||||
output as: YYYY-MM-DDTHH:MM:SSTZD (T is a literal character, TZD is
|
||||
Time Zone Designator, format +HH:MM or -HH:MM).
|
||||
|
||||
The ``HTML4()`` method below offers the same formatting, but
|
||||
converts to UTC before returning the value and sets the TZD"Z"
|
||||
|
||||
>>> dt.ISO8601()
|
||||
'1997-03-09T13:45:00-05:00'
|
||||
|
||||
|
||||
* ``HTML4()`` returns the object in the format used in the HTML4.0
|
||||
specification, one of the standard forms in ISO8601. See
|
||||
http://www.w3.org/TR/NOTE-datetime. Dates are output as:
|
||||
YYYY-MM-DDTHH:MM:SSZ (T, Z are literal characters, the time is in
|
||||
UTC.):
|
||||
|
||||
>>> dt.HTML4()
|
||||
'1997-03-09T18:45:00Z'
|
||||
|
||||
* ``JulianDay()`` returns the Julian day according to
|
||||
http://www.tondering.dk/claus/cal/node3.html#sec-calcjd
|
||||
|
||||
>>> dt.JulianDay()
|
||||
2450517
|
||||
|
||||
* ``week()`` returns the week number according to ISO
|
||||
see http://www.tondering.dk/claus/cal/node6.html#SECTION00670000000000000000
|
||||
|
||||
>>> dt.week()
|
||||
10
|
||||
|
||||
Deprecated API
|
||||
~~~~~~~~~~~~~~
|
||||
|
||||
* DayOfWeek(): see Day()
|
||||
|
||||
* Day_(): see pDay()
|
||||
|
||||
* Mon(): see aMonth()
|
||||
|
||||
* Mon_(): see pMonth
|
||||
|
||||
General Services Provided by DateTime
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
DateTimes can be repr()'ed; the result will be a string indicating how
|
||||
to make a DateTime object like this:
|
||||
|
||||
>>> repr(dt)
|
||||
"DateTime('1997/03/09 13:45:00 US/Eastern')"
|
||||
|
||||
When we convert them into a string, we get a nicer string that could
|
||||
actually be shown to a user:
|
||||
|
||||
>>> str(dt)
|
||||
'1997/03/09 13:45:00 US/Eastern'
|
||||
|
||||
The hash value of a DateTime is based on the date and time and is
|
||||
equal for different representations of the DateTime:
|
||||
|
||||
>>> hash(dt)
|
||||
3618678
|
||||
>>> hash(dt.toZone('UTC'))
|
||||
3618678
|
||||
|
||||
DateTime objects can be compared to other DateTime objects OR floating
|
||||
point numbers such as the ones which are returned by the Python time
|
||||
module by using the equalTo method. Using this API, True is returned if the
|
||||
object represents a date/time equal to the specified DateTime or time module
|
||||
style time:
|
||||
|
||||
>>> dt.equalTo(dt)
|
||||
True
|
||||
>>> dt.equalTo(dt.toZone('UTC'))
|
||||
True
|
||||
>>> dt.equalTo(dt.timeTime())
|
||||
True
|
||||
>>> dt.equalTo(DateTime())
|
||||
False
|
||||
|
||||
Same goes for inequalities:
|
||||
|
||||
>>> dt.notEqualTo(dt)
|
||||
False
|
||||
>>> dt.notEqualTo(dt.toZone('UTC'))
|
||||
False
|
||||
>>> dt.notEqualTo(dt.timeTime())
|
||||
False
|
||||
>>> dt.notEqualTo(DateTime())
|
||||
True
|
||||
|
||||
Normal equality operations only work with DateTime objects and take the
|
||||
timezone setting into account:
|
||||
|
||||
>>> dt == dt
|
||||
True
|
||||
>>> dt == dt.toZone('UTC')
|
||||
False
|
||||
>>> dt == DateTime()
|
||||
False
|
||||
|
||||
>>> dt != dt
|
||||
False
|
||||
>>> dt != dt.toZone('UTC')
|
||||
True
|
||||
>>> dt != DateTime()
|
||||
True
|
||||
|
||||
But the other comparison operations compare the referenced moment in time and
|
||||
not the representation itself:
|
||||
|
||||
>>> dt > dt
|
||||
False
|
||||
>>> DateTime() > dt
|
||||
True
|
||||
>>> dt > DateTime().timeTime()
|
||||
False
|
||||
>>> DateTime().timeTime() > dt
|
||||
True
|
||||
|
||||
>>> dt.greaterThan(dt)
|
||||
False
|
||||
>>> DateTime().greaterThan(dt)
|
||||
True
|
||||
>>> dt.greaterThan(DateTime().timeTime())
|
||||
False
|
||||
|
||||
>>> dt >= dt
|
||||
True
|
||||
>>> DateTime() >= dt
|
||||
True
|
||||
>>> dt >= DateTime().timeTime()
|
||||
False
|
||||
>>> DateTime().timeTime() >= dt
|
||||
True
|
||||
|
||||
>>> dt.greaterThanEqualTo(dt)
|
||||
True
|
||||
>>> DateTime().greaterThanEqualTo(dt)
|
||||
True
|
||||
>>> dt.greaterThanEqualTo(DateTime().timeTime())
|
||||
False
|
||||
|
||||
>>> dt < dt
|
||||
False
|
||||
>>> DateTime() < dt
|
||||
False
|
||||
>>> dt < DateTime().timeTime()
|
||||
True
|
||||
>>> DateTime().timeTime() < dt
|
||||
False
|
||||
|
||||
>>> dt.lessThan(dt)
|
||||
False
|
||||
>>> DateTime().lessThan(dt)
|
||||
False
|
||||
>>> dt.lessThan(DateTime().timeTime())
|
||||
True
|
||||
|
||||
>>> dt <= dt
|
||||
True
|
||||
>>> DateTime() <= dt
|
||||
False
|
||||
>>> dt <= DateTime().timeTime()
|
||||
True
|
||||
>>> DateTime().timeTime() <= dt
|
||||
False
|
||||
|
||||
>>> dt.lessThanEqualTo(dt)
|
||||
True
|
||||
>>> DateTime().lessThanEqualTo(dt)
|
||||
False
|
||||
>>> dt.lessThanEqualTo(DateTime().timeTime())
|
||||
True
|
||||
|
||||
Numeric Services Provided by DateTime
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
A DateTime may be added to a number and a number may be added to a
|
||||
DateTime:
|
||||
|
||||
>>> dt + 5
|
||||
DateTime('1997/03/14 13:45:00 US/Eastern')
|
||||
>>> 5 + dt
|
||||
DateTime('1997/03/14 13:45:00 US/Eastern')
|
||||
|
||||
Two DateTimes cannot be added:
|
||||
|
||||
>>> from DateTime.interfaces import DateTimeError
|
||||
>>> try:
|
||||
... dt + dt
|
||||
... print('fail')
|
||||
... except DateTimeError:
|
||||
... print('ok')
|
||||
ok
|
||||
|
||||
Either a DateTime or a number may be subtracted from a DateTime,
|
||||
however, a DateTime may not be subtracted from a number:
|
||||
|
||||
>>> DateTime('1997/03/10 13:45 US/Eastern') - dt
|
||||
1.0
|
||||
>>> dt - 1
|
||||
DateTime('1997/03/08 13:45:00 US/Eastern')
|
||||
>>> 1 - dt
|
||||
Traceback (most recent call last):
|
||||
...
|
||||
TypeError: unsupported operand type(s) for -: 'int' and 'DateTime'
|
||||
|
||||
DateTimes can also be converted to integers (number of seconds since
|
||||
the epoch) and floats:
|
||||
|
||||
>>> int(dt)
|
||||
857933100
|
||||
>>> float(dt)
|
||||
857933100.0
|
||||
18
testclass/lib/python3.12/site-packages/DateTime/__init__.py
Normal file
18
testclass/lib/python3.12/site-packages/DateTime/__init__.py
Normal file
@ -0,0 +1,18 @@
|
||||
##############################################################################
|
||||
#
|
||||
# Copyright (c) 2002 Zope Foundation and Contributors.
|
||||
#
|
||||
# This software is subject to the provisions of the Zope Public License,
|
||||
# Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution.
|
||||
# THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED
|
||||
# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS
|
||||
# FOR A PARTICULAR PURPOSE
|
||||
#
|
||||
##############################################################################
|
||||
|
||||
from .DateTime import DateTime
|
||||
from .DateTime import Timezones
|
||||
|
||||
|
||||
__all__ = ('DateTime', 'Timezones')
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
375
testclass/lib/python3.12/site-packages/DateTime/interfaces.py
Normal file
375
testclass/lib/python3.12/site-packages/DateTime/interfaces.py
Normal file
@ -0,0 +1,375 @@
|
||||
##############################################################################
|
||||
#
|
||||
# Copyright (c) 2005 Zope Foundation and Contributors.
|
||||
#
|
||||
# This software is subject to the provisions of the Zope Public License,
|
||||
# Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution.
|
||||
# THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED
|
||||
# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS
|
||||
# FOR A PARTICULAR PURPOSE
|
||||
#
|
||||
##############################################################################
|
||||
from zope.interface import Interface
|
||||
|
||||
|
||||
class DateTimeError(Exception):
|
||||
pass
|
||||
|
||||
|
||||
class SyntaxError(DateTimeError):
|
||||
pass
|
||||
|
||||
|
||||
class DateError(DateTimeError):
|
||||
pass
|
||||
|
||||
|
||||
class TimeError(DateTimeError):
|
||||
pass
|
||||
|
||||
|
||||
class IDateTime(Interface):
|
||||
# Conversion and comparison methods
|
||||
|
||||
def localZone(ltm=None):
|
||||
"""Returns the time zone on the given date. The time zone
|
||||
can change according to daylight savings."""
|
||||
|
||||
def timeTime():
|
||||
"""Return the date/time as a floating-point number in UTC, in
|
||||
the format used by the Python time module. Note that it is
|
||||
possible to create date/time values with DateTime that have no
|
||||
meaningful value to the time module."""
|
||||
|
||||
def toZone(z):
|
||||
"""Return a DateTime with the value as the current object,
|
||||
represented in the indicated timezone."""
|
||||
|
||||
def isFuture():
|
||||
"""Return true if this object represents a date/time later
|
||||
than the time of the call"""
|
||||
|
||||
def isPast():
|
||||
"""Return true if this object represents a date/time earlier
|
||||
than the time of the call"""
|
||||
|
||||
def isCurrentYear():
|
||||
"""Return true if this object represents a date/time that
|
||||
falls within the current year, in the context of this
|
||||
object's timezone representation"""
|
||||
|
||||
def isCurrentMonth():
|
||||
"""Return true if this object represents a date/time that
|
||||
falls within the current month, in the context of this
|
||||
object's timezone representation"""
|
||||
|
||||
def isCurrentDay():
|
||||
"""Return true if this object represents a date/time that
|
||||
falls within the current day, in the context of this object's
|
||||
timezone representation"""
|
||||
|
||||
def isCurrentHour():
|
||||
"""Return true if this object represents a date/time that
|
||||
falls within the current hour, in the context of this object's
|
||||
timezone representation"""
|
||||
|
||||
def isCurrentMinute():
|
||||
"""Return true if this object represents a date/time that
|
||||
falls within the current minute, in the context of this
|
||||
object's timezone representation"""
|
||||
|
||||
def isLeapYear():
|
||||
"""Return true if the current year (in the context of the
|
||||
object's timezone) is a leap year"""
|
||||
|
||||
def earliestTime():
|
||||
"""Return a new DateTime object that represents the earliest
|
||||
possible time (in whole seconds) that still falls within the
|
||||
current object's day, in the object's timezone context"""
|
||||
|
||||
def latestTime():
|
||||
"""Return a new DateTime object that represents the latest
|
||||
possible time (in whole seconds) that still falls within the
|
||||
current object's day, in the object's timezone context"""
|
||||
|
||||
def greaterThan(t):
|
||||
"""Compare this DateTime object to another DateTime object OR
|
||||
a floating point number such as that which is returned by the
|
||||
Python time module. Returns true if the object represents a
|
||||
date/time greater than the specified DateTime or time module
|
||||
style time. Revised to give more correct results through
|
||||
comparison of long integer milliseconds."""
|
||||
|
||||
__gt__ = greaterThan
|
||||
|
||||
def greaterThanEqualTo(t):
|
||||
"""Compare this DateTime object to another DateTime object OR
|
||||
a floating point number such as that which is returned by the
|
||||
Python time module. Returns true if the object represents a
|
||||
date/time greater than or equal to the specified DateTime or
|
||||
time module style time. Revised to give more correct results
|
||||
through comparison of long integer milliseconds."""
|
||||
|
||||
__ge__ = greaterThanEqualTo
|
||||
|
||||
def equalTo(t):
|
||||
"""Compare this DateTime object to another DateTime object OR
|
||||
a floating point number such as that which is returned by the
|
||||
Python time module. Returns true if the object represents a
|
||||
date/time equal to the specified DateTime or time module style
|
||||
time. Revised to give more correct results through comparison
|
||||
of long integer milliseconds."""
|
||||
|
||||
__eq__ = equalTo
|
||||
|
||||
def notEqualTo(t):
|
||||
"""Compare this DateTime object to another DateTime object OR
|
||||
a floating point number such as that which is returned by the
|
||||
Python time module. Returns true if the object represents a
|
||||
date/time not equal to the specified DateTime or time module
|
||||
style time. Revised to give more correct results through
|
||||
comparison of long integer milliseconds."""
|
||||
|
||||
__ne__ = notEqualTo
|
||||
|
||||
def lessThan(t):
|
||||
"""Compare this DateTime object to another DateTime object OR
|
||||
a floating point number such as that which is returned by the
|
||||
Python time module. Returns true if the object represents a
|
||||
date/time less than the specified DateTime or time module
|
||||
style time. Revised to give more correct results through
|
||||
comparison of long integer milliseconds."""
|
||||
|
||||
__lt__ = lessThan
|
||||
|
||||
def lessThanEqualTo(t):
|
||||
"""Compare this DateTime object to another DateTime object OR
|
||||
a floating point number such as that which is returned by the
|
||||
Python time module. Returns true if the object represents a
|
||||
date/time less than or equal to the specified DateTime or time
|
||||
module style time. Revised to give more correct results
|
||||
through comparison of long integer milliseconds."""
|
||||
|
||||
__le__ = lessThanEqualTo
|
||||
|
||||
# Component access
|
||||
|
||||
def parts():
|
||||
"""Return a tuple containing the calendar year, month, day,
|
||||
hour, minute second and timezone of the object"""
|
||||
|
||||
def timezone():
|
||||
"""Return the timezone in which the object is represented."""
|
||||
|
||||
def tzoffset():
|
||||
"""Return the timezone offset for the objects timezone."""
|
||||
|
||||
def year():
|
||||
"""Return the calendar year of the object"""
|
||||
|
||||
def month():
|
||||
"""Return the month of the object as an integer"""
|
||||
|
||||
def Month():
|
||||
"""Return the full month name"""
|
||||
|
||||
def aMonth():
|
||||
"""Return the abbreviated month name."""
|
||||
|
||||
def Mon():
|
||||
"""Compatibility: see aMonth"""
|
||||
|
||||
def pMonth():
|
||||
"""Return the abbreviated (with period) month name."""
|
||||
|
||||
def Mon_():
|
||||
"""Compatibility: see pMonth"""
|
||||
|
||||
def day():
|
||||
"""Return the integer day"""
|
||||
|
||||
def Day():
|
||||
"""Return the full name of the day of the week"""
|
||||
|
||||
def DayOfWeek():
|
||||
"""Compatibility: see Day"""
|
||||
|
||||
def dayOfYear():
|
||||
"""Return the day of the year, in context of the timezone
|
||||
representation of the object"""
|
||||
|
||||
def aDay():
|
||||
"""Return the abbreviated name of the day of the week"""
|
||||
|
||||
def pDay():
|
||||
"""Return the abbreviated (with period) name of the day of the
|
||||
week"""
|
||||
|
||||
def Day_():
|
||||
"""Compatibility: see pDay"""
|
||||
|
||||
def dow():
|
||||
"""Return the integer day of the week, where sunday is 0"""
|
||||
|
||||
def dow_1():
|
||||
"""Return the integer day of the week, where sunday is 1"""
|
||||
|
||||
def h_12():
|
||||
"""Return the 12-hour clock representation of the hour"""
|
||||
|
||||
def h_24():
|
||||
"""Return the 24-hour clock representation of the hour"""
|
||||
|
||||
def ampm():
|
||||
"""Return the appropriate time modifier (am or pm)"""
|
||||
|
||||
def hour():
|
||||
"""Return the 24-hour clock representation of the hour"""
|
||||
|
||||
def minute():
|
||||
"""Return the minute"""
|
||||
|
||||
def second():
|
||||
"""Return the second"""
|
||||
|
||||
def millis():
|
||||
"""Return the millisecond since the epoch in GMT."""
|
||||
|
||||
def strftime(format):
|
||||
"""Format the date/time using the *current timezone representation*."""
|
||||
|
||||
# General formats from previous DateTime
|
||||
|
||||
def Date():
|
||||
"""Return the date string for the object."""
|
||||
|
||||
def Time():
|
||||
"""Return the time string for an object to the nearest second."""
|
||||
|
||||
def TimeMinutes():
|
||||
"""Return the time string for an object not showing seconds."""
|
||||
|
||||
def AMPM():
|
||||
"""Return the time string for an object to the nearest second."""
|
||||
|
||||
def AMPMMinutes():
|
||||
"""Return the time string for an object not showing seconds."""
|
||||
|
||||
def PreciseTime():
|
||||
"""Return the time string for the object."""
|
||||
|
||||
def PreciseAMPM():
|
||||
"""Return the time string for the object."""
|
||||
|
||||
def yy():
|
||||
"""Return calendar year as a 2 digit string"""
|
||||
|
||||
def mm():
|
||||
"""Return month as a 2 digit string"""
|
||||
|
||||
def dd():
|
||||
"""Return day as a 2 digit string"""
|
||||
|
||||
def rfc822():
|
||||
"""Return the date in RFC 822 format"""
|
||||
|
||||
# New formats
|
||||
|
||||
def fCommon():
|
||||
"""Return a string representing the object's value in the
|
||||
format: March 1, 1997 1:45 pm"""
|
||||
|
||||
def fCommonZ():
|
||||
"""Return a string representing the object's value in the
|
||||
format: March 1, 1997 1:45 pm US/Eastern"""
|
||||
|
||||
def aCommon():
|
||||
"""Return a string representing the object's value in the
|
||||
format: Mar 1, 1997 1:45 pm"""
|
||||
|
||||
def aCommonZ():
|
||||
"""Return a string representing the object's value in the
|
||||
format: Mar 1, 1997 1:45 pm US/Eastern"""
|
||||
|
||||
def pCommon():
|
||||
"""Return a string representing the object's value in the
|
||||
format: Mar. 1, 1997 1:45 pm"""
|
||||
|
||||
def pCommonZ():
|
||||
"""Return a string representing the object's value
|
||||
in the format: Mar. 1, 1997 1:45 pm US/Eastern"""
|
||||
|
||||
def ISO():
|
||||
"""Return the object in ISO standard format. Note: this is
|
||||
*not* ISO 8601-format! See the ISO8601 and HTML4 methods below
|
||||
for ISO 8601-compliant output
|
||||
|
||||
Dates are output as: YYYY-MM-DD HH:MM:SS
|
||||
"""
|
||||
|
||||
def ISO8601():
|
||||
"""Return the object in ISO 8601-compatible format containing
|
||||
the date, time with seconds-precision and the time zone
|
||||
identifier - see http://www.w3.org/TR/NOTE-datetime
|
||||
|
||||
Dates are output as: YYYY-MM-DDTHH:MM:SSTZD
|
||||
T is a literal character.
|
||||
TZD is Time Zone Designator, format +HH:MM or -HH:MM
|
||||
|
||||
The HTML4 method below offers the same formatting, but
|
||||
converts to UTC before returning the value and sets the TZD"Z"
|
||||
"""
|
||||
|
||||
def HTML4():
|
||||
"""Return the object in the format used in the HTML4.0
|
||||
specification, one of the standard forms in ISO8601. See
|
||||
http://www.w3.org/TR/NOTE-datetime
|
||||
|
||||
Dates are output as: YYYY-MM-DDTHH:MM:SSZ
|
||||
T, Z are literal characters.
|
||||
The time is in UTC.
|
||||
"""
|
||||
|
||||
def JulianDay():
|
||||
"""Return the Julian day according to
|
||||
https://www.tondering.dk/claus/cal/julperiod.php#formula
|
||||
"""
|
||||
|
||||
def week():
|
||||
"""Return the week number according to ISO.
|
||||
|
||||
See https://www.tondering.dk/claus/cal/week.php#weekno
|
||||
"""
|
||||
|
||||
# Python operator and conversion API
|
||||
|
||||
def __add__(other):
|
||||
"""A DateTime may be added to a number and a number may be
|
||||
added to a DateTime; two DateTimes cannot be added."""
|
||||
|
||||
__radd__ = __add__
|
||||
|
||||
def __sub__(other):
|
||||
"""Either a DateTime or a number may be subtracted from a
|
||||
DateTime, however, a DateTime may not be subtracted from a
|
||||
number."""
|
||||
|
||||
def __repr__():
|
||||
"""Convert a DateTime to a string that looks like a Python
|
||||
expression."""
|
||||
|
||||
def __str__():
|
||||
"""Convert a DateTime to a string."""
|
||||
|
||||
def __hash__():
|
||||
"""Compute a hash value for a DateTime"""
|
||||
|
||||
def __int__():
|
||||
"""Convert to an integer number of seconds since the epoch (gmt)"""
|
||||
|
||||
def __long__():
|
||||
"""Convert to a long-int number of seconds since the epoch (gmt)"""
|
||||
|
||||
def __float__():
|
||||
"""Convert to floating-point number of seconds since the epoch (gmt)"""
|
||||
192
testclass/lib/python3.12/site-packages/DateTime/pytz.txt
Normal file
192
testclass/lib/python3.12/site-packages/DateTime/pytz.txt
Normal file
@ -0,0 +1,192 @@
|
||||
Pytz Support
|
||||
============
|
||||
|
||||
Allows the pytz package to be used for time zone information. The
|
||||
advantage of using pytz is that it has a more complete and up to date
|
||||
time zone and daylight savings time database.
|
||||
|
||||
Usage
|
||||
-----
|
||||
You don't have to do anything special to make it work.
|
||||
|
||||
>>> from DateTime import DateTime, Timezones
|
||||
>>> d = DateTime('March 11, 2007 US/Eastern')
|
||||
|
||||
Daylight Savings
|
||||
----------------
|
||||
In 2007 daylight savings time in the US was changed. The Energy Policy
|
||||
Act of 2005 mandates that DST will start on the second Sunday in March
|
||||
and end on the first Sunday in November.
|
||||
|
||||
In 2007, the start and stop dates are March 11 and November 4,
|
||||
respectively. These dates are different from previous DST start and
|
||||
stop dates. In 2006, the dates were the first Sunday in April (April
|
||||
2, 2006) and the last Sunday in October (October 29, 2006).
|
||||
|
||||
Let's make sure that DateTime can deal with this, since the primary
|
||||
motivation to use pytz for time zone information is the fact that it
|
||||
is kept up to date with daylight savings changes.
|
||||
|
||||
>>> DateTime('March 11, 2007 US/Eastern').tzoffset()
|
||||
-18000
|
||||
>>> DateTime('March 12, 2007 US/Eastern').tzoffset()
|
||||
-14400
|
||||
>>> DateTime('November 4, 2007 US/Eastern').tzoffset()
|
||||
-14400
|
||||
>>> DateTime('November 5, 2007 US/Eastern').tzoffset()
|
||||
-18000
|
||||
|
||||
Let's compare this to 2006.
|
||||
|
||||
>>> DateTime('April 2, 2006 US/Eastern').tzoffset()
|
||||
-18000
|
||||
>>> DateTime('April 3, 2006 US/Eastern').tzoffset()
|
||||
-14400
|
||||
>>> DateTime('October 29, 2006 US/Eastern').tzoffset()
|
||||
-14400
|
||||
>>> DateTime('October 30, 2006 US/Eastern').tzoffset()
|
||||
-18000
|
||||
|
||||
Time Zones
|
||||
---------
|
||||
DateTime can use pytz's large database of time zones. Here are some
|
||||
examples:
|
||||
|
||||
>>> d = DateTime('Pacific/Kwajalein')
|
||||
>>> d = DateTime('America/Shiprock')
|
||||
>>> d = DateTime('Africa/Ouagadougou')
|
||||
|
||||
Of course pytz doesn't know about everything.
|
||||
|
||||
>>> from DateTime.interfaces import SyntaxError
|
||||
>>> try:
|
||||
... d = DateTime('July 21, 1969 Moon/Eastern')
|
||||
... print('fail')
|
||||
... except SyntaxError:
|
||||
... print('ok')
|
||||
ok
|
||||
|
||||
You can still use zone names that DateTime defines that aren't part of
|
||||
the pytz database.
|
||||
|
||||
>>> d = DateTime('eet')
|
||||
>>> d = DateTime('iceland')
|
||||
|
||||
These time zones use DateTimes database. So it's preferable to use the
|
||||
official time zone name.
|
||||
|
||||
One trickiness is that DateTime supports some zone name
|
||||
abbreviations. Some of these map to pytz names, so these abbreviations
|
||||
will give you time zone date from pytz. Notable among abbreviations
|
||||
that work this way are 'est', 'cst', 'mst', and 'pst'.
|
||||
|
||||
Let's verify that 'est' picks up the 2007 daylight savings time changes.
|
||||
|
||||
>>> DateTime('March 11, 2007 est').tzoffset()
|
||||
-18000
|
||||
>>> DateTime('March 12, 2007 est').tzoffset()
|
||||
-14400
|
||||
>>> DateTime('November 4, 2007 est').tzoffset()
|
||||
-14400
|
||||
>>> DateTime('November 5, 2007 est').tzoffset()
|
||||
-18000
|
||||
|
||||
You can get a list of time zones supported by calling the Timezones() function.
|
||||
|
||||
>>> Timezones() #doctest: +ELLIPSIS
|
||||
['Africa/Abidjan', 'Africa/Accra', 'Africa/Addis_Ababa', ...]
|
||||
|
||||
Note that you can mess with this list without hurting things.
|
||||
|
||||
>>> t = Timezones()
|
||||
>>> t.remove('US/Eastern')
|
||||
>>> d = DateTime('US/Eastern')
|
||||
|
||||
|
||||
Internal Components
|
||||
-------------------
|
||||
|
||||
The following are tests of internal components.
|
||||
|
||||
Cache
|
||||
~~~~~
|
||||
|
||||
The DateTime class uses a new time zone cache.
|
||||
|
||||
>>> from DateTime.DateTime import _TZINFO
|
||||
>>> _TZINFO #doctest: +ELLIPSIS
|
||||
<DateTime.pytz_support.PytzCache ...>
|
||||
|
||||
The cache maps time zone names to time zone instances.
|
||||
|
||||
>>> cache = _TZINFO
|
||||
>>> tz = cache['GMT+730']
|
||||
>>> tz = cache['US/Mountain']
|
||||
|
||||
The cache also must provide a few attributes for use by the DateTime
|
||||
class.
|
||||
|
||||
The _zlst attribute is a list of supported time zone names.
|
||||
|
||||
>>> cache._zlst #doctest: +ELLIPSIS
|
||||
['Africa/Abidjan'... 'Africa/Accra'... 'IDLE'... 'NZST'... 'NZT'...]
|
||||
|
||||
The _zidx attribute is a list of lower-case and possibly abbreviated
|
||||
time zone names that can be mapped to official zone names.
|
||||
|
||||
>>> 'australia/yancowinna' in cache._zidx
|
||||
True
|
||||
>>> 'europe/isle_of_man' in cache._zidx
|
||||
True
|
||||
>>> 'gmt+0500' in cache._zidx
|
||||
True
|
||||
|
||||
Note that there are more items in _zidx than in _zlst since there are
|
||||
multiple names for some time zones.
|
||||
|
||||
>>> len(cache._zidx) > len(cache._zlst)
|
||||
True
|
||||
|
||||
Each entry in _zlst should also be present in _zidx in lower case form.
|
||||
|
||||
>>> for name in cache._zlst:
|
||||
... if not name.lower() in cache._zidx:
|
||||
... print("Error %s not in _zidx" % name.lower())
|
||||
|
||||
The _zmap attribute maps the names in _zidx to official names in _zlst.
|
||||
|
||||
>>> cache._zmap['africa/abidjan']
|
||||
'Africa/Abidjan'
|
||||
>>> cache._zmap['gmt+1']
|
||||
'GMT+1'
|
||||
>>> cache._zmap['gmt+0100']
|
||||
'GMT+1'
|
||||
>>> cache._zmap['utc']
|
||||
'UTC'
|
||||
|
||||
Let's make sure that _zmap and _zidx agree.
|
||||
|
||||
>>> idx = set(cache._zidx)
|
||||
>>> keys = set(cache._zmap.keys())
|
||||
>>> idx == keys
|
||||
True
|
||||
|
||||
Timezone objects
|
||||
~~~~~~~~~~~~~~~~
|
||||
The timezone instances have only one public method info(). It returns
|
||||
a tuple of (offset, is_dst, name). The method takes a timestamp, which
|
||||
is used to determine dst information.
|
||||
|
||||
>>> t1 = DateTime('November 4, 00:00 2007 US/Mountain').timeTime()
|
||||
>>> t2 = DateTime('November 4, 02:00 2007 US/Mountain').timeTime()
|
||||
>>> tz.info(t1)
|
||||
(-21600, 1, 'MDT')
|
||||
>>> tz.info(t2)
|
||||
(-25200, 0, 'MST')
|
||||
|
||||
If you don't pass any arguments to info it provides daylight savings
|
||||
time information as of today.
|
||||
|
||||
>>> tz.info() in ((-21600, 1, 'MDT'), (-25200, 0, 'MST'))
|
||||
True
|
||||
|
||||
269
testclass/lib/python3.12/site-packages/DateTime/pytz_support.py
Normal file
269
testclass/lib/python3.12/site-packages/DateTime/pytz_support.py
Normal file
@ -0,0 +1,269 @@
|
||||
##############################################################################
|
||||
#
|
||||
# Copyright (c) 2007 Zope Foundation and Contributors.
|
||||
#
|
||||
# This software is subject to the provisions of the Zope Public License,
|
||||
# Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution.
|
||||
# THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED
|
||||
# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS
|
||||
# FOR A PARTICULAR PURPOSE
|
||||
#
|
||||
##############################################################################
|
||||
|
||||
from datetime import datetime
|
||||
from datetime import timedelta
|
||||
|
||||
import pytz
|
||||
import pytz.reference
|
||||
from pytz.tzinfo import StaticTzInfo
|
||||
from pytz.tzinfo import memorized_timedelta
|
||||
|
||||
from .interfaces import DateTimeError
|
||||
|
||||
|
||||
EPOCH = datetime.fromtimestamp(0, tz=pytz.utc)
|
||||
|
||||
_numeric_timezone_data = {
|
||||
'GMT': ('GMT', 0, 1, [], '', [(0, 0, 0)], 'GMT\000'),
|
||||
'GMT+0': ('GMT+0', 0, 1, [], '', [(0, 0, 0)], 'GMT+0000\000'),
|
||||
'GMT+1': ('GMT+1', 0, 1, [], '', [(3600, 0, 0)], 'GMT+0100\000'),
|
||||
'GMT+2': ('GMT+2', 0, 1, [], '', [(7200, 0, 0)], 'GMT+0200\000'),
|
||||
'GMT+3': ('GMT+3', 0, 1, [], '', [(10800, 0, 0)], 'GMT+0300\000'),
|
||||
'GMT+4': ('GMT+4', 0, 1, [], '', [(14400, 0, 0)], 'GMT+0400\000'),
|
||||
'GMT+5': ('GMT+5', 0, 1, [], '', [(18000, 0, 0)], 'GMT+0500\000'),
|
||||
'GMT+6': ('GMT+6', 0, 1, [], '', [(21600, 0, 0)], 'GMT+0600\000'),
|
||||
'GMT+7': ('GMT+7', 0, 1, [], '', [(25200, 0, 0)], 'GMT+0700\000'),
|
||||
'GMT+8': ('GMT+8', 0, 1, [], '', [(28800, 0, 0)], 'GMT+0800\000'),
|
||||
'GMT+9': ('GMT+9', 0, 1, [], '', [(32400, 0, 0)], 'GMT+0900\000'),
|
||||
'GMT+10': ('GMT+10', 0, 1, [], '', [(36000, 0, 0)], 'GMT+1000\000'),
|
||||
'GMT+11': ('GMT+11', 0, 1, [], '', [(39600, 0, 0)], 'GMT+1100\000'),
|
||||
'GMT+12': ('GMT+12', 0, 1, [], '', [(43200, 0, 0)], 'GMT+1200\000'),
|
||||
'GMT+13': ('GMT+13', 0, 1, [], '', [(46800, 0, 0)], 'GMT+1300\000'),
|
||||
|
||||
'GMT-1': ('GMT-1', 0, 1, [], '', [(-3600, 0, 0)], 'GMT-0100\000'),
|
||||
'GMT-2': ('GMT-2', 0, 1, [], '', [(-7200, 0, 0)], 'GMT-0200\000'),
|
||||
'GMT-3': ('GMT-3', 0, 1, [], '', [(-10800, 0, 0)], 'GMT-0300\000'),
|
||||
'GMT-4': ('GMT-4', 0, 1, [], '', [(-14400, 0, 0)], 'GMT-0400\000'),
|
||||
'GMT-5': ('GMT-5', 0, 1, [], '', [(-18000, 0, 0)], 'GMT-0500\000'),
|
||||
'GMT-6': ('GMT-6', 0, 1, [], '', [(-21600, 0, 0)], 'GMT-0600\000'),
|
||||
'GMT-7': ('GMT-7', 0, 1, [], '', [(-25200, 0, 0)], 'GMT-0700\000'),
|
||||
'GMT-8': ('GMT-8', 0, 1, [], '', [(-28800, 0, 0)], 'GMT-0800\000'),
|
||||
'GMT-9': ('GMT-9', 0, 1, [], '', [(-32400, 0, 0)], 'GMT-0900\000'),
|
||||
'GMT-10': ('GMT-10', 0, 1, [], '', [(-36000, 0, 0)], 'GMT-1000\000'),
|
||||
'GMT-11': ('GMT-11', 0, 1, [], '', [(-39600, 0, 0)], 'GMT-1100\000'),
|
||||
'GMT-12': ('GMT-12', 0, 1, [], '', [(-43200, 0, 0)], 'GMT-1200\000'),
|
||||
|
||||
'GMT+0130': ('GMT+0130', 0, 1, [], '', [(5400, 0, 0)], 'GMT+0130\000'),
|
||||
'GMT+0230': ('GMT+0230', 0, 1, [], '', [(9000, 0, 0)], 'GMT+0230\000'),
|
||||
'GMT+0330': ('GMT+0330', 0, 1, [], '', [(12600, 0, 0)], 'GMT+0330\000'),
|
||||
'GMT+0430': ('GMT+0430', 0, 1, [], '', [(16200, 0, 0)], 'GMT+0430\000'),
|
||||
'GMT+0530': ('GMT+0530', 0, 1, [], '', [(19800, 0, 0)], 'GMT+0530\000'),
|
||||
'GMT+0630': ('GMT+0630', 0, 1, [], '', [(23400, 0, 0)], 'GMT+0630\000'),
|
||||
'GMT+0730': ('GMT+0730', 0, 1, [], '', [(27000, 0, 0)], 'GMT+0730\000'),
|
||||
'GMT+0830': ('GMT+0830', 0, 1, [], '', [(30600, 0, 0)], 'GMT+0830\000'),
|
||||
'GMT+0930': ('GMT+0930', 0, 1, [], '', [(34200, 0, 0)], 'GMT+0930\000'),
|
||||
'GMT+1030': ('GMT+1030', 0, 1, [], '', [(37800, 0, 0)], 'GMT+1030\000'),
|
||||
'GMT+1130': ('GMT+1130', 0, 1, [], '', [(41400, 0, 0)], 'GMT+1130\000'),
|
||||
'GMT+1230': ('GMT+1230', 0, 1, [], '', [(45000, 0, 0)], 'GMT+1230\000'),
|
||||
|
||||
'GMT-0130': ('GMT-0130', 0, 1, [], '', [(-5400, 0, 0)], 'GMT-0130\000'),
|
||||
'GMT-0230': ('GMT-0230', 0, 1, [], '', [(-9000, 0, 0)], 'GMT-0230\000'),
|
||||
'GMT-0330': ('GMT-0330', 0, 1, [], '', [(-12600, 0, 0)], 'GMT-0330\000'),
|
||||
'GMT-0430': ('GMT-0430', 0, 1, [], '', [(-16200, 0, 0)], 'GMT-0430\000'),
|
||||
'GMT-0530': ('GMT-0530', 0, 1, [], '', [(-19800, 0, 0)], 'GMT-0530\000'),
|
||||
'GMT-0630': ('GMT-0630', 0, 1, [], '', [(-23400, 0, 0)], 'GMT-0630\000'),
|
||||
'GMT-0730': ('GMT-0730', 0, 1, [], '', [(-27000, 0, 0)], 'GMT-0730\000'),
|
||||
'GMT-0830': ('GMT-0830', 0, 1, [], '', [(-30600, 0, 0)], 'GMT-0830\000'),
|
||||
'GMT-0930': ('GMT-0930', 0, 1, [], '', [(-34200, 0, 0)], 'GMT-0930\000'),
|
||||
'GMT-1030': ('GMT-1030', 0, 1, [], '', [(-37800, 0, 0)], 'GMT-1030\000'),
|
||||
'GMT-1130': ('GMT-1130', 0, 1, [], '', [(-41400, 0, 0)], 'GMT-1130\000'),
|
||||
'GMT-1230': ('GMT-1230', 0, 1, [], '', [(-45000, 0, 0)], 'GMT-1230\000'),
|
||||
}
|
||||
|
||||
# These are the timezones not in pytz.common_timezones
|
||||
_old_zlst = [
|
||||
'AST', 'AT', 'BST', 'BT', 'CCT',
|
||||
'CET', 'CST', 'Cuba', 'EADT', 'EAST',
|
||||
'EEST', 'EET', 'EST', 'Egypt', 'FST',
|
||||
'FWT', 'GB-Eire', 'GMT+0100', 'GMT+0130', 'GMT+0200',
|
||||
'GMT+0230', 'GMT+0300', 'GMT+0330', 'GMT+0400', 'GMT+0430',
|
||||
'GMT+0500', 'GMT+0530', 'GMT+0600', 'GMT+0630', 'GMT+0700',
|
||||
'GMT+0730', 'GMT+0800', 'GMT+0830', 'GMT+0900', 'GMT+0930',
|
||||
'GMT+1', 'GMT+1000', 'GMT+1030', 'GMT+1100', 'GMT+1130',
|
||||
'GMT+1200', 'GMT+1230', 'GMT+1300', 'GMT-0100', 'GMT-0130',
|
||||
'GMT-0200', 'GMT-0300', 'GMT-0400', 'GMT-0500', 'GMT-0600',
|
||||
'GMT-0630', 'GMT-0700', 'GMT-0730', 'GMT-0800', 'GMT-0830',
|
||||
'GMT-0900', 'GMT-0930', 'GMT-1000', 'GMT-1030', 'GMT-1100',
|
||||
'GMT-1130', 'GMT-1200', 'GMT-1230', 'GST', 'Greenwich',
|
||||
'Hongkong', 'IDLE', 'IDLW', 'Iceland', 'Iran',
|
||||
'Israel', 'JST', 'Jamaica', 'Japan', 'MEST',
|
||||
'MET', 'MEWT', 'MST', 'NT', 'NZDT',
|
||||
'NZST', 'NZT', 'PST', 'Poland', 'SST',
|
||||
'SWT', 'Singapore', 'Turkey', 'UCT', 'UT',
|
||||
'Universal', 'WADT', 'WAST', 'WAT', 'WET',
|
||||
'ZP4', 'ZP5', 'ZP6',
|
||||
]
|
||||
|
||||
_old_zmap = {
|
||||
'aest': 'GMT+10', 'aedt': 'GMT+11',
|
||||
'aus eastern standard time': 'GMT+10',
|
||||
'sydney standard time': 'GMT+10',
|
||||
'tasmania standard time': 'GMT+10',
|
||||
'e. australia standard time': 'GMT+10',
|
||||
'aus central standard time': 'GMT+0930',
|
||||
'cen. australia standard time': 'GMT+0930',
|
||||
'w. australia standard time': 'GMT+8',
|
||||
|
||||
'central europe standard time': 'GMT+1',
|
||||
'eastern standard time': 'US/Eastern',
|
||||
'us eastern standard time': 'US/Eastern',
|
||||
'central standard time': 'US/Central',
|
||||
'mountain standard time': 'US/Mountain',
|
||||
'pacific standard time': 'US/Pacific',
|
||||
'mst': 'US/Mountain', 'pst': 'US/Pacific',
|
||||
'cst': 'US/Central', 'est': 'US/Eastern',
|
||||
|
||||
'gmt+0000': 'GMT+0', 'gmt+0': 'GMT+0',
|
||||
|
||||
'gmt+0100': 'GMT+1', 'gmt+0200': 'GMT+2', 'gmt+0300': 'GMT+3',
|
||||
'gmt+0400': 'GMT+4', 'gmt+0500': 'GMT+5', 'gmt+0600': 'GMT+6',
|
||||
'gmt+0700': 'GMT+7', 'gmt+0800': 'GMT+8', 'gmt+0900': 'GMT+9',
|
||||
'gmt+1000': 'GMT+10', 'gmt+1100': 'GMT+11', 'gmt+1200': 'GMT+12',
|
||||
'gmt+1300': 'GMT+13',
|
||||
'gmt-0100': 'GMT-1', 'gmt-0200': 'GMT-2', 'gmt-0300': 'GMT-3',
|
||||
'gmt-0400': 'GMT-4', 'gmt-0500': 'GMT-5', 'gmt-0600': 'GMT-6',
|
||||
'gmt-0700': 'GMT-7', 'gmt-0800': 'GMT-8', 'gmt-0900': 'GMT-9',
|
||||
'gmt-1000': 'GMT-10', 'gmt-1100': 'GMT-11', 'gmt-1200': 'GMT-12',
|
||||
|
||||
'gmt+1': 'GMT+1', 'gmt+2': 'GMT+2', 'gmt+3': 'GMT+3',
|
||||
'gmt+4': 'GMT+4', 'gmt+5': 'GMT+5', 'gmt+6': 'GMT+6',
|
||||
'gmt+7': 'GMT+7', 'gmt+8': 'GMT+8', 'gmt+9': 'GMT+9',
|
||||
'gmt+10': 'GMT+10', 'gmt+11': 'GMT+11', 'gmt+12': 'GMT+12',
|
||||
'gmt+13': 'GMT+13',
|
||||
'gmt-1': 'GMT-1', 'gmt-2': 'GMT-2', 'gmt-3': 'GMT-3',
|
||||
'gmt-4': 'GMT-4', 'gmt-5': 'GMT-5', 'gmt-6': 'GMT-6',
|
||||
'gmt-7': 'GMT-7', 'gmt-8': 'GMT-8', 'gmt-9': 'GMT-9',
|
||||
'gmt-10': 'GMT-10', 'gmt-11': 'GMT-11', 'gmt-12': 'GMT-12',
|
||||
|
||||
'gmt+130': 'GMT+0130', 'gmt+0130': 'GMT+0130',
|
||||
'gmt+230': 'GMT+0230', 'gmt+0230': 'GMT+0230',
|
||||
'gmt+330': 'GMT+0330', 'gmt+0330': 'GMT+0330',
|
||||
'gmt+430': 'GMT+0430', 'gmt+0430': 'GMT+0430',
|
||||
'gmt+530': 'GMT+0530', 'gmt+0530': 'GMT+0530',
|
||||
'gmt+630': 'GMT+0630', 'gmt+0630': 'GMT+0630',
|
||||
'gmt+730': 'GMT+0730', 'gmt+0730': 'GMT+0730',
|
||||
'gmt+830': 'GMT+0830', 'gmt+0830': 'GMT+0830',
|
||||
'gmt+930': 'GMT+0930', 'gmt+0930': 'GMT+0930',
|
||||
'gmt+1030': 'GMT+1030',
|
||||
'gmt+1130': 'GMT+1130',
|
||||
'gmt+1230': 'GMT+1230',
|
||||
|
||||
'gmt-130': 'GMT-0130', 'gmt-0130': 'GMT-0130',
|
||||
'gmt-230': 'GMT-0230', 'gmt-0230': 'GMT-0230',
|
||||
'gmt-330': 'GMT-0330', 'gmt-0330': 'GMT-0330',
|
||||
'gmt-430': 'GMT-0430', 'gmt-0430': 'GMT-0430',
|
||||
'gmt-530': 'GMT-0530', 'gmt-0530': 'GMT-0530',
|
||||
'gmt-630': 'GMT-0630', 'gmt-0630': 'GMT-0630',
|
||||
'gmt-730': 'GMT-0730', 'gmt-0730': 'GMT-0730',
|
||||
'gmt-830': 'GMT-0830', 'gmt-0830': 'GMT-0830',
|
||||
'gmt-930': 'GMT-0930', 'gmt-0930': 'GMT-0930',
|
||||
'gmt-1030': 'GMT-1030',
|
||||
'gmt-1130': 'GMT-1130',
|
||||
'gmt-1230': 'GMT-1230',
|
||||
|
||||
'ut': 'Universal',
|
||||
'bst': 'GMT+1', 'mest': 'GMT+2', 'sst': 'GMT+2',
|
||||
'fst': 'GMT+2', 'wadt': 'GMT+8', 'eadt': 'GMT+11', 'nzdt': 'GMT+13',
|
||||
'wet': 'GMT', 'wat': 'GMT+1', 'at': 'GMT-2', 'ast': 'GMT-4',
|
||||
'nt': 'GMT-11', 'idlw': 'GMT-12', 'cet': 'GMT+1', 'cest': 'GMT+2',
|
||||
'met': 'GMT+1',
|
||||
'mewt': 'GMT+1', 'swt': 'GMT+1', 'fwt': 'GMT+1', 'eet': 'GMT+2',
|
||||
'eest': 'GMT+3',
|
||||
'bt': 'GMT+3', 'zp4': 'GMT+4', 'zp5': 'GMT+5', 'zp6': 'GMT+6',
|
||||
'wast': 'GMT+7', 'cct': 'GMT+8', 'jst': 'GMT+9', 'east': 'GMT+10',
|
||||
'gst': 'GMT+10', 'nzt': 'GMT+12', 'nzst': 'GMT+12', 'idle': 'GMT+12',
|
||||
'ret': 'GMT+4', 'ist': 'GMT+0530', 'edt': 'GMT-4',
|
||||
|
||||
}
|
||||
|
||||
|
||||
# some timezone definitions of the "-0400" are not working
|
||||
# when upgrading
|
||||
for hour in range(0, 13):
|
||||
hour = hour
|
||||
fhour = str(hour)
|
||||
if len(fhour) == 1:
|
||||
fhour = '0' + fhour
|
||||
_old_zmap['-%s00' % fhour] = 'GMT-%i' % hour
|
||||
_old_zmap['+%s00' % fhour] = 'GMT+%i' % hour
|
||||
|
||||
|
||||
def _p(zone):
|
||||
return _numeric_timezones[zone]
|
||||
|
||||
|
||||
def _static_timezone_factory(data):
|
||||
zone = data[0]
|
||||
cls = type(zone, (StaticTzInfo,), dict(
|
||||
__reduce__=lambda _: (_p, (zone, )),
|
||||
zone=zone,
|
||||
_utcoffset=memorized_timedelta(data[5][0][0]),
|
||||
_tzname=data[6][:-1])) # strip the trailing null
|
||||
return cls()
|
||||
|
||||
|
||||
_numeric_timezones = {key: _static_timezone_factory(data)
|
||||
for key, data in _numeric_timezone_data.items()}
|
||||
|
||||
|
||||
class Timezone:
|
||||
"""
|
||||
Timezone information returned by PytzCache.__getitem__
|
||||
Adapts datetime.tzinfo object to DateTime._timezone interface
|
||||
"""
|
||||
|
||||
def __init__(self, tzinfo):
|
||||
self.tzinfo = tzinfo
|
||||
|
||||
def info(self, t=None):
|
||||
if t is None:
|
||||
dt = datetime.now(tz=pytz.utc)
|
||||
else:
|
||||
# can't use utcfromtimestamp past 2038
|
||||
dt = EPOCH + timedelta(0, t)
|
||||
|
||||
# need to normalize tzinfo for the datetime to deal with
|
||||
# daylight savings time.
|
||||
normalized_dt = self.tzinfo.normalize(dt.astimezone(self.tzinfo))
|
||||
normalized_tzinfo = normalized_dt.tzinfo
|
||||
|
||||
offset = normalized_tzinfo.utcoffset(normalized_dt)
|
||||
secs = offset.days * 24 * 60 * 60 + offset.seconds
|
||||
dst = normalized_tzinfo.dst(normalized_dt)
|
||||
if dst == timedelta(0):
|
||||
is_dst = 0
|
||||
else:
|
||||
is_dst = 1
|
||||
return secs, is_dst, normalized_tzinfo.tzname(normalized_dt)
|
||||
|
||||
|
||||
class PytzCache:
|
||||
"""
|
||||
Reimplementation of the DateTime._cache class that uses for timezone info
|
||||
"""
|
||||
|
||||
_zlst = pytz.common_timezones + _old_zlst # used by DateTime.TimeZones
|
||||
_zmap = {name.lower(): name for name in pytz.all_timezones}
|
||||
_zmap.update(_old_zmap) # These must take priority
|
||||
_zidx = _zmap.keys()
|
||||
|
||||
def __getitem__(self, key):
|
||||
name = self._zmap.get(key.lower(), key) # fallback to key
|
||||
try:
|
||||
return Timezone(pytz.timezone(name))
|
||||
except pytz.UnknownTimeZoneError:
|
||||
try:
|
||||
return Timezone(_numeric_timezones[name])
|
||||
except KeyError:
|
||||
raise DateTimeError('Unrecognized timezone: %s' % key)
|
||||
@ -0,0 +1,15 @@
|
||||
##############################################################################
|
||||
#
|
||||
# Copyright (c) 2003 Zope Foundation and Contributors.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# This software is subject to the provisions of the Zope Public License,
|
||||
# Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution.
|
||||
# THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED
|
||||
# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS
|
||||
# FOR A PARTICULAR PURPOSE.
|
||||
#
|
||||
##############################################################################
|
||||
|
||||
# This file is needed to make this a package.
|
||||
Binary file not shown.
Binary file not shown.
@ -0,0 +1,57 @@
|
||||
1970-01-01 (1970, 1, 4)
|
||||
1970-01-02 (1970, 1, 5)
|
||||
1970-01-30 (1970, 5, 5)
|
||||
1970-01-31 (1970, 5, 6)
|
||||
1970-02-01 (1970, 5, 7)
|
||||
1970-02-02 (1970, 6, 1)
|
||||
1970-02-28 (1970, 9, 6)
|
||||
1970-03-01 (1970, 9, 7)
|
||||
1970-03-30 (1970, 14, 1)
|
||||
1970-03-31 (1970, 14, 2)
|
||||
1970-04-01 (1970, 14, 3)
|
||||
1970-09-30 (1970, 40, 3)
|
||||
1970-10-01 (1970, 40, 4)
|
||||
1970-10-02 (1970, 40, 5)
|
||||
1970-10-03 (1970, 40, 6)
|
||||
1970-10-04 (1970, 40, 7)
|
||||
1970-10-05 (1970, 41, 1)
|
||||
1971-01-02 (1970, 53, 6)
|
||||
1971-01-03 (1970, 53, 7)
|
||||
1971-01-04 (1971, 1, 1)
|
||||
1971-01-05 (1971, 1, 2)
|
||||
1971-12-31 (1971, 52, 5)
|
||||
1972-01-01 (1971, 52, 6)
|
||||
1972-01-02 (1971, 52, 7)
|
||||
1972-01-03 (1972, 1, 1)
|
||||
1972-01-04 (1972, 1, 2)
|
||||
1972-12-30 (1972, 52, 6)
|
||||
1972-12-31 (1972, 52, 7)
|
||||
1973-01-01 (1973, 1, 1)
|
||||
1973-01-02 (1973, 1, 2)
|
||||
1973-12-29 (1973, 52, 6)
|
||||
1973-12-30 (1973, 52, 7)
|
||||
1973-12-31 (1974, 1, 1)
|
||||
1974-01-01 (1974, 1, 2)
|
||||
1998-12-30 (1998, 53, 3)
|
||||
1998-12-31 (1998, 53, 4)
|
||||
1999-01-01 (1998, 53, 5)
|
||||
1999-01-02 (1998, 53, 6)
|
||||
1999-01-03 (1998, 53, 7)
|
||||
1999-01-04 (1999, 1, 1)
|
||||
1999-01-05 (1999, 1, 2)
|
||||
1999-12-30 (1999, 52, 4)
|
||||
1999-12-31 (1999, 52, 5)
|
||||
2000-01-01 (1999, 52, 6)
|
||||
2000-01-02 (1999, 52, 7)
|
||||
2000-01-03 (2000, 1, 1)
|
||||
2000-01-04 (2000, 1, 2)
|
||||
2000-01-05 (2000, 1, 3)
|
||||
2000-01-06 (2000, 1, 4)
|
||||
2000-01-07 (2000, 1, 5)
|
||||
2000-01-08 (2000, 1, 6)
|
||||
2000-01-09 (2000, 1, 7)
|
||||
2000-01-10 (2000, 2, 1)
|
||||
2019-12-28 (2019, 52, 6)
|
||||
2019-12-29 (2019, 52, 7)
|
||||
2019-12-30 (2020, 1, 1)
|
||||
2019-12-31 (2020, 1, 2)
|
||||
@ -0,0 +1,764 @@
|
||||
##############################################################################
|
||||
#
|
||||
# Copyright (c) 2003 Zope Foundation and Contributors.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# This software is subject to the provisions of the Zope Public License,
|
||||
# Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution.
|
||||
# THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED
|
||||
# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS
|
||||
# FOR A PARTICULAR PURPOSE.
|
||||
#
|
||||
##############################################################################
|
||||
|
||||
import math
|
||||
import os
|
||||
import pickle
|
||||
import platform
|
||||
import sys
|
||||
import time
|
||||
import unittest
|
||||
from datetime import date
|
||||
from datetime import datetime
|
||||
from datetime import timedelta
|
||||
from datetime import tzinfo
|
||||
|
||||
import pytz
|
||||
|
||||
from DateTime import DateTime
|
||||
from DateTime.DateTime import _findLocalTimeZoneName
|
||||
|
||||
|
||||
try:
|
||||
__file__
|
||||
except NameError: # pragma: no cover
|
||||
f = sys.argv[0]
|
||||
else:
|
||||
f = __file__
|
||||
|
||||
IS_PYPY = getattr(platform, 'python_implementation', lambda: None)() == 'PyPy'
|
||||
|
||||
DATADIR = os.path.dirname(os.path.abspath(f))
|
||||
del f
|
||||
|
||||
ZERO = timedelta(0)
|
||||
|
||||
|
||||
class FixedOffset(tzinfo):
|
||||
"""Fixed offset in minutes east from UTC."""
|
||||
|
||||
def __init__(self, offset, name):
|
||||
self.__offset = timedelta(minutes=offset)
|
||||
self.__name = name
|
||||
|
||||
def utcoffset(self, dt):
|
||||
return self.__offset
|
||||
|
||||
def tzname(self, dt):
|
||||
return self.__name
|
||||
|
||||
def dst(self, dt):
|
||||
return ZERO
|
||||
|
||||
|
||||
class DateTimeTests(unittest.TestCase):
|
||||
|
||||
def _compare(self, dt1, dt2):
|
||||
'''Compares the internal representation of dt1 with
|
||||
the representation in dt2. Allows sub-millisecond variations.
|
||||
Primarily for testing.'''
|
||||
self.assertEqual(round(dt1._t, 3), round(dt2._t, 3))
|
||||
self.assertEqual(round(dt1._d, 9), round(dt2._d, 9))
|
||||
self.assertEqual(round(dt1.time, 9), round(dt2.time, 9))
|
||||
self.assertEqual(dt1.millis(), dt2.millis())
|
||||
self.assertEqual(dt1._micros, dt2._micros)
|
||||
|
||||
def testBug1203(self):
|
||||
# 01:59:60 occurred in old DateTime
|
||||
dt = DateTime(7200, 'GMT')
|
||||
self.assertTrue(str(dt).find('60') < 0, dt)
|
||||
|
||||
def testDSTInEffect(self):
|
||||
# Checks GMT offset for a DST date in the US/Eastern time zone
|
||||
dt = DateTime(2000, 5, 9, 15, 0, 0, 'US/Eastern')
|
||||
self.assertEqual(dt.toZone('GMT').hour(), 19,
|
||||
(dt, dt.toZone('GMT')))
|
||||
|
||||
def testDSTNotInEffect(self):
|
||||
# Checks GMT offset for a non-DST date in the US/Eastern time zone
|
||||
dt = DateTime(2000, 11, 9, 15, 0, 0, 'US/Eastern')
|
||||
self.assertEqual(dt.toZone('GMT').hour(), 20,
|
||||
(dt, dt.toZone('GMT')))
|
||||
|
||||
def testAddPrecision(self):
|
||||
# Precision of serial additions
|
||||
dt = DateTime()
|
||||
self.assertEqual(str(dt + 0.10 + 3.14 + 6.76 - 10), str(dt),
|
||||
dt)
|
||||
# checks problem reported in
|
||||
# https://github.com/zopefoundation/DateTime/issues/41
|
||||
dt = DateTime(2038, 10, 7, 8, 52, 44.959840, "UTC")
|
||||
self.assertEqual(str(dt + 0.10 + 3.14 + 6.76 - 10), str(dt),
|
||||
dt)
|
||||
|
||||
def testConsistentSecondMicroRounding(self):
|
||||
dt = DateTime(2038, 10, 7, 8, 52, 44.9598398, "UTC")
|
||||
self.assertEqual(int(dt.second() * 1000000),
|
||||
dt.micros() % 60000000)
|
||||
|
||||
def testConstructor3(self):
|
||||
# Constructor from date/time string
|
||||
dt = DateTime()
|
||||
dt1s = '%d/%d/%d %d:%d:%f %s' % (
|
||||
dt.year(),
|
||||
dt.month(),
|
||||
dt.day(),
|
||||
dt.hour(),
|
||||
dt.minute(),
|
||||
dt.second(),
|
||||
dt.timezone())
|
||||
dt1 = DateTime(dt1s)
|
||||
# Compare representations as it's the
|
||||
# only way to compare the dates to the same accuracy
|
||||
self.assertEqual(repr(dt), repr(dt1))
|
||||
|
||||
def testConstructor4(self):
|
||||
# Constructor from time float
|
||||
dt = DateTime()
|
||||
dt1 = DateTime(float(dt))
|
||||
self._compare(dt, dt1)
|
||||
|
||||
def testConstructor5(self):
|
||||
# Constructor from time float and timezone
|
||||
dt = DateTime()
|
||||
dt1 = DateTime(float(dt), dt.timezone())
|
||||
self.assertEqual(str(dt), str(dt1), (dt, dt1))
|
||||
dt1 = DateTime(float(dt), str(dt.timezone()))
|
||||
self.assertEqual(str(dt), str(dt1), (dt, dt1))
|
||||
|
||||
def testConstructor6(self):
|
||||
# Constructor from year and julian date
|
||||
# This test must normalize the time zone, or it *will* break when
|
||||
# DST changes!
|
||||
dt1 = DateTime(2000, 5.500000578705)
|
||||
dt = DateTime('2000/1/5 12:00:00.050 pm %s' % dt1.localZone())
|
||||
self._compare(dt, dt1)
|
||||
|
||||
def testConstructor7(self):
|
||||
# Constructor from parts
|
||||
dt = DateTime()
|
||||
dt1 = DateTime(
|
||||
dt.year(),
|
||||
dt.month(),
|
||||
dt.day(),
|
||||
dt.hour(),
|
||||
dt.minute(),
|
||||
dt.second(),
|
||||
dt.timezone())
|
||||
# Compare representations as it's the
|
||||
# only way to compare the dates to the same accuracy
|
||||
self.assertEqual(repr(dt), repr(dt1))
|
||||
|
||||
def testDayOfWeek(self):
|
||||
# Compare to the datetime.date value to make it locale independent
|
||||
expected = date(2000, 6, 16).strftime('%A')
|
||||
# strftime() used to always be passed a day of week of 0
|
||||
dt = DateTime('2000/6/16')
|
||||
s = dt.strftime('%A')
|
||||
self.assertEqual(s, expected, (dt, s))
|
||||
|
||||
def testOldDate(self):
|
||||
# Fails when an 1800 date is displayed with negative signs
|
||||
dt = DateTime('1830/5/6 12:31:46.213 pm')
|
||||
dt1 = dt.toZone('GMT+6')
|
||||
self.assertTrue(str(dt1).find('-') < 0, (dt, dt1))
|
||||
|
||||
def testSubtraction(self):
|
||||
# Reconstruction of a DateTime from its parts, with subtraction
|
||||
# this also tests the accuracy of addition and reconstruction
|
||||
dt = DateTime()
|
||||
dt1 = dt - 3.141592653
|
||||
dt2 = DateTime(
|
||||
dt.year(),
|
||||
dt.month(),
|
||||
dt.day(),
|
||||
dt.hour(),
|
||||
dt.minute(),
|
||||
dt.second())
|
||||
dt3 = dt2 - 3.141592653
|
||||
self.assertEqual(dt1, dt3, (dt, dt1, dt2, dt3))
|
||||
|
||||
def testTZ1add(self):
|
||||
# Time zone manipulation: add to a date
|
||||
dt = DateTime('1997/3/8 1:45am GMT-4')
|
||||
dt1 = DateTime('1997/3/9 1:45pm GMT+8')
|
||||
self.assertTrue((dt + 1.0).equalTo(dt1))
|
||||
|
||||
def testTZ1sub(self):
|
||||
# Time zone manipulation: subtract from a date
|
||||
dt = DateTime('1997/3/8 1:45am GMT-4')
|
||||
dt1 = DateTime('1997/3/9 1:45pm GMT+8')
|
||||
self.assertTrue((dt1 - 1.0).equalTo(dt))
|
||||
|
||||
def testTZ1diff(self):
|
||||
# Time zone manipulation: diff two dates
|
||||
dt = DateTime('1997/3/8 1:45am GMT-4')
|
||||
dt1 = DateTime('1997/3/9 1:45pm GMT+8')
|
||||
self.assertEqual(dt1 - dt, 1.0, (dt, dt1))
|
||||
|
||||
def test_compare_methods(self):
|
||||
# Compare two dates using several methods
|
||||
dt = DateTime('1997/1/1')
|
||||
dt1 = DateTime('1997/2/2')
|
||||
self.assertTrue(dt1.greaterThan(dt))
|
||||
self.assertTrue(dt1.greaterThanEqualTo(dt))
|
||||
self.assertTrue(dt.lessThan(dt1))
|
||||
self.assertTrue(dt.lessThanEqualTo(dt1))
|
||||
self.assertTrue(dt.notEqualTo(dt1))
|
||||
self.assertFalse(dt.equalTo(dt1))
|
||||
# Compare a date to float
|
||||
dt = DateTime(1.0)
|
||||
self.assertTrue(dt == DateTime(1.0)) # testing __eq__
|
||||
self.assertFalse(dt != DateTime(1.0)) # testing __ne__
|
||||
self.assertFalse(dt.greaterThan(1.0))
|
||||
self.assertTrue(dt.greaterThanEqualTo(1.0))
|
||||
self.assertFalse(dt.lessThan(1.0))
|
||||
self.assertTrue(dt.lessThanEqualTo(1.0))
|
||||
self.assertFalse(dt.notEqualTo(1.0))
|
||||
self.assertTrue(dt.equalTo(1.0))
|
||||
# Compare a date to int
|
||||
dt = DateTime(1)
|
||||
self.assertEqual(dt, DateTime(1.0))
|
||||
self.assertTrue(dt == DateTime(1)) # testing __eq__
|
||||
self.assertFalse(dt != DateTime(1)) # testing __ne__
|
||||
self.assertFalse(dt.greaterThan(1))
|
||||
self.assertTrue(dt.greaterThanEqualTo(1))
|
||||
self.assertFalse(dt.lessThan(1))
|
||||
self.assertTrue(dt.lessThanEqualTo(1))
|
||||
self.assertFalse(dt.notEqualTo(1))
|
||||
self.assertTrue(dt.equalTo(1))
|
||||
# Compare a date to string; there is no implicit type conversion
|
||||
# but behavior if consistent as when comparing, for example, an int
|
||||
# and a string.
|
||||
dt = DateTime("2023")
|
||||
self.assertFalse(dt == "2023") # testing __eq__
|
||||
self.assertTrue(dt != "2023") # testing __ne__
|
||||
self.assertRaises(TypeError, dt.greaterThan, "2023")
|
||||
self.assertRaises(TypeError, dt.greaterThanEqualTo, "2023")
|
||||
self.assertRaises(TypeError, dt.lessThan, "2023")
|
||||
self.assertRaises(TypeError, dt.lessThanEqualTo, "2023")
|
||||
self.assertTrue(dt.notEqualTo("2023"))
|
||||
self.assertFalse(dt.equalTo("2023"))
|
||||
|
||||
def test_compare_methods_none(self):
|
||||
# Compare a date to None
|
||||
for dt in (DateTime('1997/1/1'), DateTime(0)):
|
||||
self.assertTrue(dt.greaterThan(None))
|
||||
self.assertTrue(dt.greaterThanEqualTo(None))
|
||||
self.assertFalse(dt.lessThan(None))
|
||||
self.assertFalse(dt.lessThanEqualTo(None))
|
||||
self.assertTrue(dt.notEqualTo(None))
|
||||
self.assertFalse(dt.equalTo(None))
|
||||
|
||||
def test_pickle(self):
|
||||
dt = DateTime()
|
||||
data = pickle.dumps(dt, 1)
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_with_tz(self):
|
||||
dt = DateTime('2002/5/2 8:00am GMT+8')
|
||||
data = pickle.dumps(dt, 1)
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_asdatetime_with_tz(self):
|
||||
dt = DateTime('2002/5/2 8:00am GMT+8')
|
||||
data = pickle.dumps(dt.asdatetime(), 1)
|
||||
new = DateTime(pickle.loads(data))
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_with_numerical_tz(self):
|
||||
for dt_str in ('2007/01/02 12:34:56.789 +0300',
|
||||
'2007/01/02 12:34:56.789 +0430',
|
||||
'2007/01/02 12:34:56.789 -1234'):
|
||||
dt = DateTime(dt_str)
|
||||
data = pickle.dumps(dt, 1)
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_with_micros(self):
|
||||
dt = DateTime('2002/5/2 8:00:14.123 GMT+8')
|
||||
data = pickle.dumps(dt, 1)
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_old(self):
|
||||
dt = DateTime('2002/5/2 8:00am GMT+0')
|
||||
data = (
|
||||
'(cDateTime.DateTime\nDateTime\nq\x01Noq\x02}q\x03(U\x05'
|
||||
'_amonq\x04U\x03Mayq\x05U\x05_adayq\x06U\x03Thuq\x07U\x05_pmonq'
|
||||
'\x08h\x05U\x05_hourq\tK\x08U\x05_fmonq\nh\x05U\x05_pdayq\x0bU'
|
||||
'\x04Thu.q\x0cU\x05_fdayq\rU\x08Thursdayq\x0eU\x03_pmq\x0fU\x02amq'
|
||||
'\x10U\x02_tq\x11GA\xcehy\x00\x00\x00\x00U\x07_minuteq\x12K\x00U'
|
||||
'\x07_microsq\x13L1020326400000000L\nU\x02_dq\x14G@\xe2\x12j\xaa'
|
||||
'\xaa\xaa\xabU\x07_secondq\x15G\x00\x00\x00\x00\x00\x00\x00\x00U'
|
||||
'\x03_tzq\x16U\x05GMT+0q\x17U\x06_monthq\x18K\x05U'
|
||||
'\x0f_timezone_naiveq\x19I00\nU\x04_dayq\x1aK\x02U\x05_yearq'
|
||||
'\x1bM\xd2\x07U\x08_nearsecq\x1cG\x00\x00\x00\x00\x00\x00\x00'
|
||||
'\x00U\x07_pmhourq\x1dK\x08U\n_dayoffsetq\x1eK\x04U\x04timeq'
|
||||
'\x1fG?\xd5UUUV\x00\x00ub.')
|
||||
data = data.encode('latin-1')
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_old_without_micros(self):
|
||||
dt = DateTime('2002/5/2 8:00am GMT+0')
|
||||
data = (
|
||||
'(cDateTime.DateTime\nDateTime\nq\x01Noq\x02}q\x03(U\x05'
|
||||
'_amonq\x04U\x03Mayq\x05U\x05_adayq\x06U\x03Thuq\x07U\x05_pmonq'
|
||||
'\x08h\x05U\x05_hourq\tK\x08U\x05_fmonq\nh\x05U\x05_pdayq\x0bU'
|
||||
'\x04Thu.q\x0cU\x05_fdayq\rU\x08Thursdayq\x0eU\x03_pmq\x0fU'
|
||||
'\x02amq\x10U\x02_tq\x11GA\xcehy\x00\x00\x00\x00U\x07_minuteq'
|
||||
'\x12K\x00U\x02_dq\x13G@\xe2\x12j\xaa\xaa\xaa\xabU\x07_secondq'
|
||||
'\x14G\x00\x00\x00\x00\x00\x00\x00\x00U\x03_tzq\x15U\x05GMT+0q'
|
||||
'\x16U\x06_monthq\x17K\x05U\x0f_timezone_naiveq\x18I00\nU'
|
||||
'\x04_dayq\x19K\x02U\x05_yearq\x1aM\xd2\x07U\x08_nearsecq'
|
||||
'\x1bG\x00\x00\x00\x00\x00\x00\x00\x00U\x07_pmhourq\x1cK\x08U'
|
||||
'\n_dayoffsetq\x1dK\x04U\x04timeq\x1eG?\xd5UUUV\x00\x00ub.')
|
||||
data = data.encode('latin-1')
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_dates_after_2038(self):
|
||||
dt = DateTime('2039/09/02 07:07:6.235027 GMT+1')
|
||||
data = pickle.dumps(dt, 1)
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def test_pickle_old_with_micros_as_float(self):
|
||||
dt = DateTime('2002/5/2 8:00am GMT+0')
|
||||
data = (
|
||||
'ccopy_reg\n_reconstructor\nq\x00(cDateTime.DateTime\nDateTime'
|
||||
'\nq\x01c__builtin__\nobject\nq\x02Ntq\x03Rq\x04(GA\xcehy\x00\x00'
|
||||
'\x00\x00I00\nX\x05\x00\x00\x00GMT+0q\x05tq\x06b.')
|
||||
data = data.encode('latin-1')
|
||||
new = pickle.loads(data)
|
||||
for key in DateTime.__slots__:
|
||||
self.assertEqual(getattr(dt, key), getattr(new, key))
|
||||
|
||||
def testTZ2(self):
|
||||
# Time zone manipulation test 2
|
||||
dt = DateTime()
|
||||
dt1 = dt.toZone('GMT')
|
||||
s = dt.second()
|
||||
s1 = dt1.second()
|
||||
self.assertEqual(s, s1, (dt, dt1, s, s1))
|
||||
|
||||
def testTZDiffDaylight(self):
|
||||
# Diff dates across daylight savings dates
|
||||
dt = DateTime('2000/6/8 1:45am US/Eastern')
|
||||
dt1 = DateTime('2000/12/8 12:45am US/Eastern')
|
||||
self.assertEqual(dt1 - dt, 183, (dt, dt1, dt1 - dt))
|
||||
|
||||
def testY10KDate(self):
|
||||
# Comparison of a Y10K date and a Y2K date
|
||||
dt = DateTime('10213/09/21')
|
||||
dt1 = DateTime(2000, 1, 1)
|
||||
|
||||
dsec = (dt.millis() - dt1.millis()) / 1000.0
|
||||
ddays = math.floor((dsec / 86400.0) + 0.5)
|
||||
|
||||
self.assertEqual(ddays, 3000000, ddays)
|
||||
|
||||
def test_tzoffset(self):
|
||||
# Test time-zone given as an offset
|
||||
|
||||
# GMT
|
||||
dt = DateTime('Tue, 10 Sep 2001 09:41:03 GMT')
|
||||
self.assertEqual(dt.tzoffset(), 0)
|
||||
|
||||
# Timezone by name, a timezone that hasn't got daylightsaving.
|
||||
dt = DateTime('Tue, 2 Mar 2001 09:41:03 GMT+3')
|
||||
self.assertEqual(dt.tzoffset(), 10800)
|
||||
|
||||
# Timezone by name, has daylightsaving but is not in effect.
|
||||
dt = DateTime('Tue, 21 Jan 2001 09:41:03 PST')
|
||||
self.assertEqual(dt.tzoffset(), -28800)
|
||||
|
||||
# Timezone by name, with daylightsaving in effect
|
||||
dt = DateTime('Tue, 24 Aug 2001 09:41:03 PST')
|
||||
self.assertEqual(dt.tzoffset(), -25200)
|
||||
|
||||
# A negative numerical timezone
|
||||
dt = DateTime('Tue, 24 Jul 2001 09:41:03 -0400')
|
||||
self.assertEqual(dt.tzoffset(), -14400)
|
||||
|
||||
# A positive numerical timzone
|
||||
dt = DateTime('Tue, 6 Dec 1966 01:41:03 +0200')
|
||||
self.assertEqual(dt.tzoffset(), 7200)
|
||||
|
||||
# A negative numerical timezone with minutes.
|
||||
dt = DateTime('Tue, 24 Jul 2001 09:41:03 -0637')
|
||||
self.assertEqual(dt.tzoffset(), -23820)
|
||||
|
||||
# A positive numerical timezone with minutes.
|
||||
dt = DateTime('Tue, 24 Jul 2001 09:41:03 +0425')
|
||||
self.assertEqual(dt.tzoffset(), 15900)
|
||||
|
||||
def testISO8601(self):
|
||||
# ISO8601 reference dates
|
||||
ref0 = DateTime('2002/5/2 8:00am GMT')
|
||||
ref1 = DateTime('2002/5/2 8:00am US/Eastern')
|
||||
ref2 = DateTime('2006/11/6 10:30 GMT')
|
||||
ref3 = DateTime('2004/06/14 14:30:15 GMT-3')
|
||||
ref4 = DateTime('2006/01/01 GMT')
|
||||
|
||||
# Basic tests
|
||||
# Though this is timezone naive and according to specification should
|
||||
# be interpreted in the local timezone, to preserve backwards
|
||||
# compatibility with previously expected behaviour.
|
||||
isoDt = DateTime('2002-05-02T08:00:00')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002-05-02T08:00:00Z')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002-05-02T08:00:00+00:00')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002-05-02T08:00:00-04:00')
|
||||
self.assertTrue(ref1.equalTo(isoDt))
|
||||
isoDt = DateTime('2002-05-02 08:00:00-04:00')
|
||||
self.assertTrue(ref1.equalTo(isoDt))
|
||||
|
||||
# Bug 1386: the colon in the timezone offset is optional
|
||||
isoDt = DateTime('2002-05-02T08:00:00-0400')
|
||||
self.assertTrue(ref1.equalTo(isoDt))
|
||||
|
||||
# Bug 2191: date reduced formats
|
||||
isoDt = DateTime('2006-01-01')
|
||||
self.assertTrue(ref4.equalTo(isoDt))
|
||||
isoDt = DateTime('200601-01')
|
||||
self.assertTrue(ref4.equalTo(isoDt))
|
||||
isoDt = DateTime('20060101')
|
||||
self.assertTrue(ref4.equalTo(isoDt))
|
||||
isoDt = DateTime('2006-01')
|
||||
self.assertTrue(ref4.equalTo(isoDt))
|
||||
isoDt = DateTime('200601')
|
||||
self.assertTrue(ref4.equalTo(isoDt))
|
||||
isoDt = DateTime('2006')
|
||||
self.assertTrue(ref4.equalTo(isoDt))
|
||||
|
||||
# Bug 2191: date/time separators are also optional
|
||||
isoDt = DateTime('20020502T08:00:00')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002-05-02T080000')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('20020502T080000')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
|
||||
# Bug 2191: timezones with only one digit for hour
|
||||
isoDt = DateTime('20020502T080000+0')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('20020502 080000-4')
|
||||
self.assertTrue(ref1.equalTo(isoDt))
|
||||
isoDt = DateTime('20020502T080000-400')
|
||||
self.assertTrue(ref1.equalTo(isoDt))
|
||||
isoDt = DateTime('20020502T080000-4:00')
|
||||
self.assertTrue(ref1.equalTo(isoDt))
|
||||
|
||||
# Bug 2191: optional seconds/minutes
|
||||
isoDt = DateTime('2002-05-02T0800')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002-05-02T08')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
|
||||
# Bug 2191: week format
|
||||
isoDt = DateTime('2002-W18-4T0800')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002-W184T0800')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002W18-4T0800')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002W184T08')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2004-W25-1T14:30:15-03:00')
|
||||
self.assertTrue(ref3.equalTo(isoDt))
|
||||
isoDt = DateTime('2004-W25T14:30:15-03:00')
|
||||
self.assertTrue(ref3.equalTo(isoDt))
|
||||
|
||||
# Bug 2191: day of year format
|
||||
isoDt = DateTime('2002-122T0800')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
isoDt = DateTime('2002122T0800')
|
||||
self.assertTrue(ref0.equalTo(isoDt))
|
||||
|
||||
# Bug 2191: hours/minutes fractions
|
||||
isoDt = DateTime('2006-11-06T10.5')
|
||||
self.assertTrue(ref2.equalTo(isoDt))
|
||||
isoDt = DateTime('2006-11-06T10,5')
|
||||
self.assertTrue(ref2.equalTo(isoDt))
|
||||
isoDt = DateTime('20040614T1430.25-3')
|
||||
self.assertTrue(ref3.equalTo(isoDt))
|
||||
isoDt = DateTime('2004-06-14T1430,25-3')
|
||||
self.assertTrue(ref3.equalTo(isoDt))
|
||||
isoDt = DateTime('2004-06-14T14:30.25-3')
|
||||
self.assertTrue(ref3.equalTo(isoDt))
|
||||
isoDt = DateTime('20040614T14:30,25-3')
|
||||
self.assertTrue(ref3.equalTo(isoDt))
|
||||
|
||||
# ISO8601 standard format
|
||||
iso8601_string = '2002-05-02T08:00:00-04:00'
|
||||
iso8601DT = DateTime(iso8601_string)
|
||||
self.assertEqual(iso8601_string, iso8601DT.ISO8601())
|
||||
|
||||
# ISO format with no timezone
|
||||
isoDt = DateTime('2006-01-01 00:00:00')
|
||||
self.assertTrue(ref4.equalTo(isoDt))
|
||||
|
||||
def testJulianWeek(self):
|
||||
# Check JulianDayWeek function
|
||||
fn = os.path.join(DATADIR, 'julian_testdata.txt')
|
||||
with open(fn) as fd:
|
||||
lines = fd.readlines()
|
||||
for line in lines:
|
||||
d = DateTime(line[:10])
|
||||
result_from_mx = tuple(map(int, line[12:-2].split(',')))
|
||||
self.assertEqual(result_from_mx[1], d.week())
|
||||
|
||||
def testCopyConstructor(self):
|
||||
d = DateTime('2004/04/04')
|
||||
self.assertEqual(DateTime(d), d)
|
||||
self.assertEqual(str(DateTime(d)), str(d))
|
||||
d2 = DateTime('1999/04/12 01:00:00')
|
||||
self.assertEqual(DateTime(d2), d2)
|
||||
self.assertEqual(str(DateTime(d2)), str(d2))
|
||||
|
||||
def testCopyConstructorPreservesTimezone(self):
|
||||
# test for https://bugs.launchpad.net/zope2/+bug/200007
|
||||
# This always worked in the local timezone, so we need at least
|
||||
# two tests with different zones to be sure at least one of them
|
||||
# is not local.
|
||||
d = DateTime('2004/04/04')
|
||||
self.assertEqual(DateTime(d).timezone(), d.timezone())
|
||||
d2 = DateTime('2008/04/25 12:00:00 EST')
|
||||
self.assertEqual(DateTime(d2).timezone(), d2.timezone())
|
||||
self.assertEqual(str(DateTime(d2)), str(d2))
|
||||
d3 = DateTime('2008/04/25 12:00:00 PST')
|
||||
self.assertEqual(DateTime(d3).timezone(), d3.timezone())
|
||||
self.assertEqual(str(DateTime(d3)), str(d3))
|
||||
|
||||
def testRFC822(self):
|
||||
# rfc822 conversion
|
||||
dt = DateTime('2002-05-02T08:00:00+00:00')
|
||||
self.assertEqual(dt.rfc822(), 'Thu, 02 May 2002 08:00:00 +0000')
|
||||
|
||||
dt = DateTime('2002-05-02T08:00:00+02:00')
|
||||
self.assertEqual(dt.rfc822(), 'Thu, 02 May 2002 08:00:00 +0200')
|
||||
|
||||
dt = DateTime('2002-05-02T08:00:00-02:00')
|
||||
self.assertEqual(dt.rfc822(), 'Thu, 02 May 2002 08:00:00 -0200')
|
||||
|
||||
# Checking that conversion from local time is working.
|
||||
dt = DateTime()
|
||||
dts = dt.rfc822().split(' ')
|
||||
times = dts[4].split(':')
|
||||
_isDST = time.localtime(time.time())[8]
|
||||
if _isDST:
|
||||
offset = time.altzone
|
||||
else:
|
||||
offset = time.timezone
|
||||
self.assertEqual(dts[0], dt.aDay() + ',')
|
||||
self.assertEqual(int(dts[1]), dt.day())
|
||||
self.assertEqual(dts[2], dt.aMonth())
|
||||
self.assertEqual(int(dts[3]), dt.year())
|
||||
self.assertEqual(int(times[0]), dt.h_24())
|
||||
self.assertEqual(int(times[1]), dt.minute())
|
||||
self.assertEqual(int(times[2]), int(dt.second()))
|
||||
self.assertEqual(dts[5], "%+03d%02d" % divmod((-offset / 60), 60))
|
||||
|
||||
def testInternationalDateformat(self):
|
||||
for year in (1990, 2001, 2020):
|
||||
for month in (1, 12):
|
||||
for day in (1, 12, 28, 31):
|
||||
try:
|
||||
d_us = DateTime("%d/%d/%d" % (year, month, day))
|
||||
except Exception:
|
||||
continue
|
||||
|
||||
d_int = DateTime("%d.%d.%d" % (day, month, year),
|
||||
datefmt="international")
|
||||
self.assertEqual(d_us, d_int)
|
||||
|
||||
d_int = DateTime("%d/%d/%d" % (day, month, year),
|
||||
datefmt="international")
|
||||
self.assertEqual(d_us, d_int)
|
||||
|
||||
def test_intl_format_hyphen(self):
|
||||
d_jan = DateTime('2011-01-11 GMT')
|
||||
d_nov = DateTime('2011-11-01 GMT')
|
||||
d_us = DateTime('11-01-2011 GMT')
|
||||
d_int = DateTime('11-01-2011 GMT', datefmt="international")
|
||||
self.assertNotEqual(d_us, d_int)
|
||||
self.assertEqual(d_us, d_nov)
|
||||
self.assertEqual(d_int, d_jan)
|
||||
|
||||
def test_calcTimezoneName(self):
|
||||
from DateTime.interfaces import TimeError
|
||||
timezone_dependent_epoch = 2177452800
|
||||
try:
|
||||
DateTime()._calcTimezoneName(timezone_dependent_epoch, 0)
|
||||
except TimeError:
|
||||
self.fail('Zope Collector issue #484 (negative time bug): '
|
||||
'TimeError raised')
|
||||
|
||||
def testStrftimeTZhandling(self):
|
||||
# strftime timezone testing
|
||||
# This is a test for collector issue #1127
|
||||
format = '%Y-%m-%d %H:%M %Z'
|
||||
dt = DateTime('Wed, 19 Nov 2003 18:32:07 -0215')
|
||||
dt_string = dt.strftime(format)
|
||||
dt_local = dt.toZone(_findLocalTimeZoneName(0))
|
||||
dt_localstring = dt_local.strftime(format)
|
||||
self.assertEqual(dt_string, dt_localstring)
|
||||
|
||||
def testStrftimeFarDates(self):
|
||||
# Checks strftime in dates <= 1900 or >= 2038
|
||||
dt = DateTime('1900/01/30')
|
||||
self.assertEqual(dt.strftime('%d/%m/%Y'), '30/01/1900')
|
||||
dt = DateTime('2040/01/30')
|
||||
self.assertEqual(dt.strftime('%d/%m/%Y'), '30/01/2040')
|
||||
|
||||
def testZoneInFarDates(self):
|
||||
# Checks time zone in dates <= 1900 or >= 2038
|
||||
dt1 = DateTime('2040/01/30 14:33 GMT+1')
|
||||
dt2 = DateTime('2040/01/30 11:33 GMT-2')
|
||||
self.assertEqual(dt1.strftime('%d/%m/%Y %H:%M'),
|
||||
dt2.strftime('%d/%m/%Y %H:%M'))
|
||||
|
||||
@unittest.skipIf(
|
||||
IS_PYPY,
|
||||
"Using Non-Ascii characters for strftime doesn't work in PyPy"
|
||||
"https://bitbucket.org/pypy/pypy/issues/2161/pypy3-strftime-does-not-accept-unicode" # noqa: E501 line too long
|
||||
)
|
||||
def testStrftimeStr(self):
|
||||
dt = DateTime('2002-05-02T08:00:00+00:00')
|
||||
uchar = b'\xc3\xa0'.decode('utf-8')
|
||||
ok = dt.strftime('Le %d/%m/%Y a %Hh%M').replace('a', uchar)
|
||||
ustr = b'Le %d/%m/%Y \xc3\xa0 %Hh%M'.decode('utf-8')
|
||||
self.assertEqual(dt.strftime(ustr), ok)
|
||||
|
||||
def testTimezoneNaiveHandling(self):
|
||||
# checks that we assign timezone naivity correctly
|
||||
dt = DateTime('2007-10-04T08:00:00+00:00')
|
||||
self.assertFalse(dt.timezoneNaive(),
|
||||
'error with naivity handling in __parse_iso8601')
|
||||
dt = DateTime('2007-10-04T08:00:00Z')
|
||||
self.assertFalse(dt.timezoneNaive(),
|
||||
'error with naivity handling in __parse_iso8601')
|
||||
dt = DateTime('2007-10-04T08:00:00')
|
||||
self.assertTrue(dt.timezoneNaive(),
|
||||
'error with naivity handling in __parse_iso8601')
|
||||
dt = DateTime('2007/10/04 15:12:33.487618 GMT+1')
|
||||
self.assertFalse(dt.timezoneNaive(),
|
||||
'error with naivity handling in _parse')
|
||||
dt = DateTime('2007/10/04 15:12:33.487618')
|
||||
self.assertTrue(dt.timezoneNaive(),
|
||||
'error with naivity handling in _parse')
|
||||
dt = DateTime()
|
||||
self.assertFalse(dt.timezoneNaive(),
|
||||
'error with naivity for current time')
|
||||
s = '2007-10-04T08:00:00'
|
||||
dt = DateTime(s)
|
||||
self.assertEqual(s, dt.ISO8601())
|
||||
s = '2007-10-04T08:00:00+00:00'
|
||||
dt = DateTime(s)
|
||||
self.assertEqual(s, dt.ISO8601())
|
||||
|
||||
def testConversions(self):
|
||||
sdt0 = datetime.now() # this is a timezone naive datetime
|
||||
dt0 = DateTime(sdt0)
|
||||
self.assertTrue(dt0.timezoneNaive(), (sdt0, dt0))
|
||||
sdt1 = datetime(2007, 10, 4, 18, 14, 42, 580, pytz.utc)
|
||||
dt1 = DateTime(sdt1)
|
||||
self.assertFalse(dt1.timezoneNaive(), (sdt1, dt1))
|
||||
|
||||
# convert back
|
||||
sdt2 = dt0.asdatetime()
|
||||
self.assertEqual(sdt0, sdt2)
|
||||
sdt3 = dt1.utcdatetime() # this returns a timezone naive datetime
|
||||
self.assertEqual(sdt1.hour, sdt3.hour)
|
||||
|
||||
dt4 = DateTime('2007-10-04T10:00:00+05:00')
|
||||
sdt4 = datetime(2007, 10, 4, 5, 0)
|
||||
self.assertEqual(dt4.utcdatetime(), sdt4)
|
||||
self.assertEqual(dt4.asdatetime(), sdt4.replace(tzinfo=pytz.utc))
|
||||
|
||||
dt5 = DateTime('2007-10-23 10:00:00 US/Eastern')
|
||||
tz = pytz.timezone('US/Eastern')
|
||||
sdt5 = datetime(2007, 10, 23, 10, 0, tzinfo=tz)
|
||||
dt6 = DateTime(sdt5)
|
||||
self.assertEqual(dt5.asdatetime(), sdt5)
|
||||
self.assertEqual(dt6.asdatetime(), sdt5)
|
||||
self.assertEqual(dt5, dt6)
|
||||
self.assertEqual(dt5.asdatetime().tzinfo, tz)
|
||||
self.assertEqual(dt6.asdatetime().tzinfo, tz)
|
||||
|
||||
def testBasicTZ(self):
|
||||
# psycopg2 supplies it's own tzinfo instances, with no `zone` attribute
|
||||
tz = FixedOffset(60, 'GMT+1')
|
||||
dt1 = datetime(2008, 8, 5, 12, 0, tzinfo=tz)
|
||||
DT = DateTime(dt1)
|
||||
dt2 = DT.asdatetime()
|
||||
offset1 = dt1.tzinfo.utcoffset(dt1)
|
||||
offset2 = dt2.tzinfo.utcoffset(dt2)
|
||||
self.assertEqual(offset1, offset2)
|
||||
|
||||
def testEDTTimezone(self):
|
||||
# should be able to parse EDT timezones: see lp:599856.
|
||||
dt = DateTime("Mon, 28 Jun 2010 10:12:25 EDT")
|
||||
self.assertEqual(dt.Day(), 'Monday')
|
||||
self.assertEqual(dt.day(), 28)
|
||||
self.assertEqual(dt.Month(), 'June')
|
||||
self.assertEqual(dt.timezone(), 'GMT-4')
|
||||
|
||||
def testParseISO8601(self):
|
||||
parsed = DateTime()._parse_iso8601('2010-10-10')
|
||||
self.assertEqual(parsed, (2010, 10, 10, 0, 0, 0, 'GMT+0000'))
|
||||
|
||||
def test_interface(self):
|
||||
from DateTime.interfaces import IDateTime
|
||||
self.assertTrue(IDateTime.providedBy(DateTime()))
|
||||
|
||||
def test_security(self):
|
||||
dt = DateTime()
|
||||
self.assertEqual(dt.__roles__, None)
|
||||
self.assertEqual(dt.__allow_access_to_unprotected_subobjects__, 1)
|
||||
|
||||
def test_format(self):
|
||||
dt = DateTime(1968, 3, 10, 23, 45, 0, 'Europe/Vienna')
|
||||
fmt = '%d.%m.%Y %H:%M'
|
||||
result = dt.strftime(fmt)
|
||||
unformatted_result = '1968/03/10 23:45:00 Europe/Vienna'
|
||||
self.assertEqual(result, f'{dt:%d.%m.%Y %H:%M}')
|
||||
self.assertEqual(unformatted_result, f'{dt}')
|
||||
self.assertEqual(unformatted_result, f'{dt}')
|
||||
self.assertEqual(result, f'{dt:{fmt}}')
|
||||
self.assertEqual(unformatted_result, f'{dt:}')
|
||||
self.assertEqual(unformatted_result, f'{dt}')
|
||||
|
||||
|
||||
def test_suite():
|
||||
import doctest
|
||||
return unittest.TestSuite([
|
||||
unittest.defaultTestLoader.loadTestsFromTestCase(DateTimeTests),
|
||||
doctest.DocFileSuite('DateTime.txt', package='DateTime'),
|
||||
doctest.DocFileSuite('pytz.txt', package='DateTime'),
|
||||
])
|
||||
133
testclass/lib/python3.12/site-packages/PIL/BdfFontFile.py
Normal file
133
testclass/lib/python3.12/site-packages/PIL/BdfFontFile.py
Normal file
@ -0,0 +1,133 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# bitmap distribution font (bdf) file parser
|
||||
#
|
||||
# history:
|
||||
# 1996-05-16 fl created (as bdf2pil)
|
||||
# 1997-08-25 fl converted to FontFile driver
|
||||
# 2001-05-25 fl removed bogus __init__ call
|
||||
# 2002-11-20 fl robustification (from Kevin Cazabon, Dmitry Vasiliev)
|
||||
# 2003-04-22 fl more robustification (from Graham Dumpleton)
|
||||
#
|
||||
# Copyright (c) 1997-2003 by Secret Labs AB.
|
||||
# Copyright (c) 1997-2003 by Fredrik Lundh.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
|
||||
"""
|
||||
Parse X Bitmap Distribution Format (BDF)
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import BinaryIO
|
||||
|
||||
from . import FontFile, Image
|
||||
|
||||
bdf_slant = {
|
||||
"R": "Roman",
|
||||
"I": "Italic",
|
||||
"O": "Oblique",
|
||||
"RI": "Reverse Italic",
|
||||
"RO": "Reverse Oblique",
|
||||
"OT": "Other",
|
||||
}
|
||||
|
||||
bdf_spacing = {"P": "Proportional", "M": "Monospaced", "C": "Cell"}
|
||||
|
||||
|
||||
def bdf_char(
|
||||
f: BinaryIO,
|
||||
) -> (
|
||||
tuple[
|
||||
str,
|
||||
int,
|
||||
tuple[tuple[int, int], tuple[int, int, int, int], tuple[int, int, int, int]],
|
||||
Image.Image,
|
||||
]
|
||||
| None
|
||||
):
|
||||
# skip to STARTCHAR
|
||||
while True:
|
||||
s = f.readline()
|
||||
if not s:
|
||||
return None
|
||||
if s[:9] == b"STARTCHAR":
|
||||
break
|
||||
id = s[9:].strip().decode("ascii")
|
||||
|
||||
# load symbol properties
|
||||
props = {}
|
||||
while True:
|
||||
s = f.readline()
|
||||
if not s or s[:6] == b"BITMAP":
|
||||
break
|
||||
i = s.find(b" ")
|
||||
props[s[:i].decode("ascii")] = s[i + 1 : -1].decode("ascii")
|
||||
|
||||
# load bitmap
|
||||
bitmap = bytearray()
|
||||
while True:
|
||||
s = f.readline()
|
||||
if not s or s[:7] == b"ENDCHAR":
|
||||
break
|
||||
bitmap += s[:-1]
|
||||
|
||||
# The word BBX
|
||||
# followed by the width in x (BBw), height in y (BBh),
|
||||
# and x and y displacement (BBxoff0, BByoff0)
|
||||
# of the lower left corner from the origin of the character.
|
||||
width, height, x_disp, y_disp = (int(p) for p in props["BBX"].split())
|
||||
|
||||
# The word DWIDTH
|
||||
# followed by the width in x and y of the character in device pixels.
|
||||
dwx, dwy = (int(p) for p in props["DWIDTH"].split())
|
||||
|
||||
bbox = (
|
||||
(dwx, dwy),
|
||||
(x_disp, -y_disp - height, width + x_disp, -y_disp),
|
||||
(0, 0, width, height),
|
||||
)
|
||||
|
||||
try:
|
||||
im = Image.frombytes("1", (width, height), bitmap, "hex", "1")
|
||||
except ValueError:
|
||||
# deal with zero-width characters
|
||||
im = Image.new("1", (width, height))
|
||||
|
||||
return id, int(props["ENCODING"]), bbox, im
|
||||
|
||||
|
||||
class BdfFontFile(FontFile.FontFile):
|
||||
"""Font file plugin for the X11 BDF format."""
|
||||
|
||||
def __init__(self, fp: BinaryIO) -> None:
|
||||
super().__init__()
|
||||
|
||||
s = fp.readline()
|
||||
if s[:13] != b"STARTFONT 2.1":
|
||||
msg = "not a valid BDF file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
props = {}
|
||||
comments = []
|
||||
|
||||
while True:
|
||||
s = fp.readline()
|
||||
if not s or s[:13] == b"ENDPROPERTIES":
|
||||
break
|
||||
i = s.find(b" ")
|
||||
props[s[:i].decode("ascii")] = s[i + 1 : -1].decode("ascii")
|
||||
if s[:i] in [b"COMMENT", b"COPYRIGHT"]:
|
||||
if s.find(b"LogicalFontDescription") < 0:
|
||||
comments.append(s[i + 1 : -1].decode("ascii"))
|
||||
|
||||
while True:
|
||||
c = bdf_char(fp)
|
||||
if not c:
|
||||
break
|
||||
id, ch, (xy, dst, src), im = c
|
||||
if 0 <= ch < len(self.glyph):
|
||||
self.glyph[ch] = xy, dst, src, im
|
||||
501
testclass/lib/python3.12/site-packages/PIL/BlpImagePlugin.py
Normal file
501
testclass/lib/python3.12/site-packages/PIL/BlpImagePlugin.py
Normal file
@ -0,0 +1,501 @@
|
||||
"""
|
||||
Blizzard Mipmap Format (.blp)
|
||||
Jerome Leclanche <jerome@leclan.ch>
|
||||
|
||||
The contents of this file are hereby released in the public domain (CC0)
|
||||
Full text of the CC0 license:
|
||||
https://creativecommons.org/publicdomain/zero/1.0/
|
||||
|
||||
BLP1 files, used mostly in Warcraft III, are not fully supported.
|
||||
All types of BLP2 files used in World of Warcraft are supported.
|
||||
|
||||
The BLP file structure consists of a header, up to 16 mipmaps of the
|
||||
texture
|
||||
|
||||
Texture sizes must be powers of two, though the two dimensions do
|
||||
not have to be equal; 512x256 is valid, but 512x200 is not.
|
||||
The first mipmap (mipmap #0) is the full size image; each subsequent
|
||||
mipmap halves both dimensions. The final mipmap should be 1x1.
|
||||
|
||||
BLP files come in many different flavours:
|
||||
* JPEG-compressed (type == 0) - only supported for BLP1.
|
||||
* RAW images (type == 1, encoding == 1). Each mipmap is stored as an
|
||||
array of 8-bit values, one per pixel, left to right, top to bottom.
|
||||
Each value is an index to the palette.
|
||||
* DXT-compressed (type == 1, encoding == 2):
|
||||
- DXT1 compression is used if alpha_encoding == 0.
|
||||
- An additional alpha bit is used if alpha_depth == 1.
|
||||
- DXT3 compression is used if alpha_encoding == 1.
|
||||
- DXT5 compression is used if alpha_encoding == 7.
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import abc
|
||||
import os
|
||||
import struct
|
||||
from enum import IntEnum
|
||||
from io import BytesIO
|
||||
from typing import IO
|
||||
|
||||
from . import Image, ImageFile
|
||||
|
||||
|
||||
class Format(IntEnum):
|
||||
JPEG = 0
|
||||
|
||||
|
||||
class Encoding(IntEnum):
|
||||
UNCOMPRESSED = 1
|
||||
DXT = 2
|
||||
UNCOMPRESSED_RAW_BGRA = 3
|
||||
|
||||
|
||||
class AlphaEncoding(IntEnum):
|
||||
DXT1 = 0
|
||||
DXT3 = 1
|
||||
DXT5 = 7
|
||||
|
||||
|
||||
def unpack_565(i: int) -> tuple[int, int, int]:
|
||||
return ((i >> 11) & 0x1F) << 3, ((i >> 5) & 0x3F) << 2, (i & 0x1F) << 3
|
||||
|
||||
|
||||
def decode_dxt1(
|
||||
data: bytes, alpha: bool = False
|
||||
) -> tuple[bytearray, bytearray, bytearray, bytearray]:
|
||||
"""
|
||||
input: one "row" of data (i.e. will produce 4*width pixels)
|
||||
"""
|
||||
|
||||
blocks = len(data) // 8 # number of blocks in row
|
||||
ret = (bytearray(), bytearray(), bytearray(), bytearray())
|
||||
|
||||
for block_index in range(blocks):
|
||||
# Decode next 8-byte block.
|
||||
idx = block_index * 8
|
||||
color0, color1, bits = struct.unpack_from("<HHI", data, idx)
|
||||
|
||||
r0, g0, b0 = unpack_565(color0)
|
||||
r1, g1, b1 = unpack_565(color1)
|
||||
|
||||
# Decode this block into 4x4 pixels
|
||||
# Accumulate the results onto our 4 row accumulators
|
||||
for j in range(4):
|
||||
for i in range(4):
|
||||
# get next control op and generate a pixel
|
||||
|
||||
control = bits & 3
|
||||
bits = bits >> 2
|
||||
|
||||
a = 0xFF
|
||||
if control == 0:
|
||||
r, g, b = r0, g0, b0
|
||||
elif control == 1:
|
||||
r, g, b = r1, g1, b1
|
||||
elif control == 2:
|
||||
if color0 > color1:
|
||||
r = (2 * r0 + r1) // 3
|
||||
g = (2 * g0 + g1) // 3
|
||||
b = (2 * b0 + b1) // 3
|
||||
else:
|
||||
r = (r0 + r1) // 2
|
||||
g = (g0 + g1) // 2
|
||||
b = (b0 + b1) // 2
|
||||
elif control == 3:
|
||||
if color0 > color1:
|
||||
r = (2 * r1 + r0) // 3
|
||||
g = (2 * g1 + g0) // 3
|
||||
b = (2 * b1 + b0) // 3
|
||||
else:
|
||||
r, g, b, a = 0, 0, 0, 0
|
||||
|
||||
if alpha:
|
||||
ret[j].extend([r, g, b, a])
|
||||
else:
|
||||
ret[j].extend([r, g, b])
|
||||
|
||||
return ret
|
||||
|
||||
|
||||
def decode_dxt3(data: bytes) -> tuple[bytearray, bytearray, bytearray, bytearray]:
|
||||
"""
|
||||
input: one "row" of data (i.e. will produce 4*width pixels)
|
||||
"""
|
||||
|
||||
blocks = len(data) // 16 # number of blocks in row
|
||||
ret = (bytearray(), bytearray(), bytearray(), bytearray())
|
||||
|
||||
for block_index in range(blocks):
|
||||
idx = block_index * 16
|
||||
block = data[idx : idx + 16]
|
||||
# Decode next 16-byte block.
|
||||
bits = struct.unpack_from("<8B", block)
|
||||
color0, color1 = struct.unpack_from("<HH", block, 8)
|
||||
|
||||
(code,) = struct.unpack_from("<I", block, 12)
|
||||
|
||||
r0, g0, b0 = unpack_565(color0)
|
||||
r1, g1, b1 = unpack_565(color1)
|
||||
|
||||
for j in range(4):
|
||||
high = False # Do we want the higher bits?
|
||||
for i in range(4):
|
||||
alphacode_index = (4 * j + i) // 2
|
||||
a = bits[alphacode_index]
|
||||
if high:
|
||||
high = False
|
||||
a >>= 4
|
||||
else:
|
||||
high = True
|
||||
a &= 0xF
|
||||
a *= 17 # We get a value between 0 and 15
|
||||
|
||||
color_code = (code >> 2 * (4 * j + i)) & 0x03
|
||||
|
||||
if color_code == 0:
|
||||
r, g, b = r0, g0, b0
|
||||
elif color_code == 1:
|
||||
r, g, b = r1, g1, b1
|
||||
elif color_code == 2:
|
||||
r = (2 * r0 + r1) // 3
|
||||
g = (2 * g0 + g1) // 3
|
||||
b = (2 * b0 + b1) // 3
|
||||
elif color_code == 3:
|
||||
r = (2 * r1 + r0) // 3
|
||||
g = (2 * g1 + g0) // 3
|
||||
b = (2 * b1 + b0) // 3
|
||||
|
||||
ret[j].extend([r, g, b, a])
|
||||
|
||||
return ret
|
||||
|
||||
|
||||
def decode_dxt5(data: bytes) -> tuple[bytearray, bytearray, bytearray, bytearray]:
|
||||
"""
|
||||
input: one "row" of data (i.e. will produce 4 * width pixels)
|
||||
"""
|
||||
|
||||
blocks = len(data) // 16 # number of blocks in row
|
||||
ret = (bytearray(), bytearray(), bytearray(), bytearray())
|
||||
|
||||
for block_index in range(blocks):
|
||||
idx = block_index * 16
|
||||
block = data[idx : idx + 16]
|
||||
# Decode next 16-byte block.
|
||||
a0, a1 = struct.unpack_from("<BB", block)
|
||||
|
||||
bits = struct.unpack_from("<6B", block, 2)
|
||||
alphacode1 = bits[2] | (bits[3] << 8) | (bits[4] << 16) | (bits[5] << 24)
|
||||
alphacode2 = bits[0] | (bits[1] << 8)
|
||||
|
||||
color0, color1 = struct.unpack_from("<HH", block, 8)
|
||||
|
||||
(code,) = struct.unpack_from("<I", block, 12)
|
||||
|
||||
r0, g0, b0 = unpack_565(color0)
|
||||
r1, g1, b1 = unpack_565(color1)
|
||||
|
||||
for j in range(4):
|
||||
for i in range(4):
|
||||
# get next control op and generate a pixel
|
||||
alphacode_index = 3 * (4 * j + i)
|
||||
|
||||
if alphacode_index <= 12:
|
||||
alphacode = (alphacode2 >> alphacode_index) & 0x07
|
||||
elif alphacode_index == 15:
|
||||
alphacode = (alphacode2 >> 15) | ((alphacode1 << 1) & 0x06)
|
||||
else: # alphacode_index >= 18 and alphacode_index <= 45
|
||||
alphacode = (alphacode1 >> (alphacode_index - 16)) & 0x07
|
||||
|
||||
if alphacode == 0:
|
||||
a = a0
|
||||
elif alphacode == 1:
|
||||
a = a1
|
||||
elif a0 > a1:
|
||||
a = ((8 - alphacode) * a0 + (alphacode - 1) * a1) // 7
|
||||
elif alphacode == 6:
|
||||
a = 0
|
||||
elif alphacode == 7:
|
||||
a = 255
|
||||
else:
|
||||
a = ((6 - alphacode) * a0 + (alphacode - 1) * a1) // 5
|
||||
|
||||
color_code = (code >> 2 * (4 * j + i)) & 0x03
|
||||
|
||||
if color_code == 0:
|
||||
r, g, b = r0, g0, b0
|
||||
elif color_code == 1:
|
||||
r, g, b = r1, g1, b1
|
||||
elif color_code == 2:
|
||||
r = (2 * r0 + r1) // 3
|
||||
g = (2 * g0 + g1) // 3
|
||||
b = (2 * b0 + b1) // 3
|
||||
elif color_code == 3:
|
||||
r = (2 * r1 + r0) // 3
|
||||
g = (2 * g1 + g0) // 3
|
||||
b = (2 * b1 + b0) // 3
|
||||
|
||||
ret[j].extend([r, g, b, a])
|
||||
|
||||
return ret
|
||||
|
||||
|
||||
class BLPFormatError(NotImplementedError):
|
||||
pass
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] in (b"BLP1", b"BLP2")
|
||||
|
||||
|
||||
class BlpImageFile(ImageFile.ImageFile):
|
||||
"""
|
||||
Blizzard Mipmap Format
|
||||
"""
|
||||
|
||||
format = "BLP"
|
||||
format_description = "Blizzard Mipmap Format"
|
||||
|
||||
def _open(self) -> None:
|
||||
self.magic = self.fp.read(4)
|
||||
if not _accept(self.magic):
|
||||
msg = f"Bad BLP magic {repr(self.magic)}"
|
||||
raise BLPFormatError(msg)
|
||||
|
||||
compression = struct.unpack("<i", self.fp.read(4))[0]
|
||||
if self.magic == b"BLP1":
|
||||
alpha = struct.unpack("<I", self.fp.read(4))[0] != 0
|
||||
else:
|
||||
encoding = struct.unpack("<b", self.fp.read(1))[0]
|
||||
alpha = struct.unpack("<b", self.fp.read(1))[0] != 0
|
||||
alpha_encoding = struct.unpack("<b", self.fp.read(1))[0]
|
||||
self.fp.seek(1, os.SEEK_CUR) # mips
|
||||
|
||||
self._size = struct.unpack("<II", self.fp.read(8))
|
||||
|
||||
args: tuple[int, int, bool] | tuple[int, int, bool, int]
|
||||
if self.magic == b"BLP1":
|
||||
encoding = struct.unpack("<i", self.fp.read(4))[0]
|
||||
self.fp.seek(4, os.SEEK_CUR) # subtype
|
||||
|
||||
args = (compression, encoding, alpha)
|
||||
offset = 28
|
||||
else:
|
||||
args = (compression, encoding, alpha, alpha_encoding)
|
||||
offset = 20
|
||||
|
||||
decoder = self.magic.decode()
|
||||
|
||||
self._mode = "RGBA" if alpha else "RGB"
|
||||
self.tile = [ImageFile._Tile(decoder, (0, 0) + self.size, offset, args)]
|
||||
|
||||
|
||||
class _BLPBaseDecoder(ImageFile.PyDecoder):
|
||||
_pulls_fd = True
|
||||
|
||||
def decode(self, buffer: bytes | Image.SupportsArrayInterface) -> tuple[int, int]:
|
||||
try:
|
||||
self._read_header()
|
||||
self._load()
|
||||
except struct.error as e:
|
||||
msg = "Truncated BLP file"
|
||||
raise OSError(msg) from e
|
||||
return -1, 0
|
||||
|
||||
@abc.abstractmethod
|
||||
def _load(self) -> None:
|
||||
pass
|
||||
|
||||
def _read_header(self) -> None:
|
||||
self._offsets = struct.unpack("<16I", self._safe_read(16 * 4))
|
||||
self._lengths = struct.unpack("<16I", self._safe_read(16 * 4))
|
||||
|
||||
def _safe_read(self, length: int) -> bytes:
|
||||
assert self.fd is not None
|
||||
return ImageFile._safe_read(self.fd, length)
|
||||
|
||||
def _read_palette(self) -> list[tuple[int, int, int, int]]:
|
||||
ret = []
|
||||
for i in range(256):
|
||||
try:
|
||||
b, g, r, a = struct.unpack("<4B", self._safe_read(4))
|
||||
except struct.error:
|
||||
break
|
||||
ret.append((b, g, r, a))
|
||||
return ret
|
||||
|
||||
def _read_bgra(
|
||||
self, palette: list[tuple[int, int, int, int]], alpha: bool
|
||||
) -> bytearray:
|
||||
data = bytearray()
|
||||
_data = BytesIO(self._safe_read(self._lengths[0]))
|
||||
while True:
|
||||
try:
|
||||
(offset,) = struct.unpack("<B", _data.read(1))
|
||||
except struct.error:
|
||||
break
|
||||
b, g, r, a = palette[offset]
|
||||
d: tuple[int, ...] = (r, g, b)
|
||||
if alpha:
|
||||
d += (a,)
|
||||
data.extend(d)
|
||||
return data
|
||||
|
||||
|
||||
class BLP1Decoder(_BLPBaseDecoder):
|
||||
def _load(self) -> None:
|
||||
self._compression, self._encoding, alpha = self.args
|
||||
|
||||
if self._compression == Format.JPEG:
|
||||
self._decode_jpeg_stream()
|
||||
|
||||
elif self._compression == 1:
|
||||
if self._encoding in (4, 5):
|
||||
palette = self._read_palette()
|
||||
data = self._read_bgra(palette, alpha)
|
||||
self.set_as_raw(data)
|
||||
else:
|
||||
msg = f"Unsupported BLP encoding {repr(self._encoding)}"
|
||||
raise BLPFormatError(msg)
|
||||
else:
|
||||
msg = f"Unsupported BLP compression {repr(self._encoding)}"
|
||||
raise BLPFormatError(msg)
|
||||
|
||||
def _decode_jpeg_stream(self) -> None:
|
||||
from .JpegImagePlugin import JpegImageFile
|
||||
|
||||
(jpeg_header_size,) = struct.unpack("<I", self._safe_read(4))
|
||||
jpeg_header = self._safe_read(jpeg_header_size)
|
||||
assert self.fd is not None
|
||||
self._safe_read(self._offsets[0] - self.fd.tell()) # What IS this?
|
||||
data = self._safe_read(self._lengths[0])
|
||||
data = jpeg_header + data
|
||||
image = JpegImageFile(BytesIO(data))
|
||||
Image._decompression_bomb_check(image.size)
|
||||
if image.mode == "CMYK":
|
||||
decoder_name, extents, offset, args = image.tile[0]
|
||||
assert isinstance(args, tuple)
|
||||
image.tile = [
|
||||
ImageFile._Tile(decoder_name, extents, offset, (args[0], "CMYK"))
|
||||
]
|
||||
r, g, b = image.convert("RGB").split()
|
||||
reversed_image = Image.merge("RGB", (b, g, r))
|
||||
self.set_as_raw(reversed_image.tobytes())
|
||||
|
||||
|
||||
class BLP2Decoder(_BLPBaseDecoder):
|
||||
def _load(self) -> None:
|
||||
self._compression, self._encoding, alpha, self._alpha_encoding = self.args
|
||||
|
||||
palette = self._read_palette()
|
||||
|
||||
assert self.fd is not None
|
||||
self.fd.seek(self._offsets[0])
|
||||
|
||||
if self._compression == 1:
|
||||
# Uncompressed or DirectX compression
|
||||
|
||||
if self._encoding == Encoding.UNCOMPRESSED:
|
||||
data = self._read_bgra(palette, alpha)
|
||||
|
||||
elif self._encoding == Encoding.DXT:
|
||||
data = bytearray()
|
||||
if self._alpha_encoding == AlphaEncoding.DXT1:
|
||||
linesize = (self.state.xsize + 3) // 4 * 8
|
||||
for yb in range((self.state.ysize + 3) // 4):
|
||||
for d in decode_dxt1(self._safe_read(linesize), alpha):
|
||||
data += d
|
||||
|
||||
elif self._alpha_encoding == AlphaEncoding.DXT3:
|
||||
linesize = (self.state.xsize + 3) // 4 * 16
|
||||
for yb in range((self.state.ysize + 3) // 4):
|
||||
for d in decode_dxt3(self._safe_read(linesize)):
|
||||
data += d
|
||||
|
||||
elif self._alpha_encoding == AlphaEncoding.DXT5:
|
||||
linesize = (self.state.xsize + 3) // 4 * 16
|
||||
for yb in range((self.state.ysize + 3) // 4):
|
||||
for d in decode_dxt5(self._safe_read(linesize)):
|
||||
data += d
|
||||
else:
|
||||
msg = f"Unsupported alpha encoding {repr(self._alpha_encoding)}"
|
||||
raise BLPFormatError(msg)
|
||||
else:
|
||||
msg = f"Unknown BLP encoding {repr(self._encoding)}"
|
||||
raise BLPFormatError(msg)
|
||||
|
||||
else:
|
||||
msg = f"Unknown BLP compression {repr(self._compression)}"
|
||||
raise BLPFormatError(msg)
|
||||
|
||||
self.set_as_raw(data)
|
||||
|
||||
|
||||
class BLPEncoder(ImageFile.PyEncoder):
|
||||
_pushes_fd = True
|
||||
|
||||
def _write_palette(self) -> bytes:
|
||||
data = b""
|
||||
assert self.im is not None
|
||||
palette = self.im.getpalette("RGBA", "RGBA")
|
||||
for i in range(len(palette) // 4):
|
||||
r, g, b, a = palette[i * 4 : (i + 1) * 4]
|
||||
data += struct.pack("<4B", b, g, r, a)
|
||||
while len(data) < 256 * 4:
|
||||
data += b"\x00" * 4
|
||||
return data
|
||||
|
||||
def encode(self, bufsize: int) -> tuple[int, int, bytes]:
|
||||
palette_data = self._write_palette()
|
||||
|
||||
offset = 20 + 16 * 4 * 2 + len(palette_data)
|
||||
data = struct.pack("<16I", offset, *((0,) * 15))
|
||||
|
||||
assert self.im is not None
|
||||
w, h = self.im.size
|
||||
data += struct.pack("<16I", w * h, *((0,) * 15))
|
||||
|
||||
data += palette_data
|
||||
|
||||
for y in range(h):
|
||||
for x in range(w):
|
||||
data += struct.pack("<B", self.im.getpixel((x, y)))
|
||||
|
||||
return len(data), 0, data
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
if im.mode != "P":
|
||||
msg = "Unsupported BLP image mode"
|
||||
raise ValueError(msg)
|
||||
|
||||
magic = b"BLP1" if im.encoderinfo.get("blp_version") == "BLP1" else b"BLP2"
|
||||
fp.write(magic)
|
||||
|
||||
assert im.palette is not None
|
||||
fp.write(struct.pack("<i", 1)) # Uncompressed or DirectX compression
|
||||
|
||||
alpha_depth = 1 if im.palette.mode == "RGBA" else 0
|
||||
if magic == b"BLP1":
|
||||
fp.write(struct.pack("<L", alpha_depth))
|
||||
else:
|
||||
fp.write(struct.pack("<b", Encoding.UNCOMPRESSED))
|
||||
fp.write(struct.pack("<b", alpha_depth))
|
||||
fp.write(struct.pack("<b", 0)) # alpha encoding
|
||||
fp.write(struct.pack("<b", 0)) # mips
|
||||
fp.write(struct.pack("<II", *im.size))
|
||||
if magic == b"BLP1":
|
||||
fp.write(struct.pack("<i", 5))
|
||||
fp.write(struct.pack("<i", 0))
|
||||
|
||||
ImageFile._save(im, fp, [ImageFile._Tile("BLP", (0, 0) + im.size, 0, im.mode)])
|
||||
|
||||
|
||||
Image.register_open(BlpImageFile.format, BlpImageFile, _accept)
|
||||
Image.register_extension(BlpImageFile.format, ".blp")
|
||||
Image.register_decoder("BLP1", BLP1Decoder)
|
||||
Image.register_decoder("BLP2", BLP2Decoder)
|
||||
|
||||
Image.register_save(BlpImageFile.format, _save)
|
||||
Image.register_encoder("BLP", BLPEncoder)
|
||||
511
testclass/lib/python3.12/site-packages/PIL/BmpImagePlugin.py
Normal file
511
testclass/lib/python3.12/site-packages/PIL/BmpImagePlugin.py
Normal file
@ -0,0 +1,511 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# BMP file handler
|
||||
#
|
||||
# Windows (and OS/2) native bitmap storage format.
|
||||
#
|
||||
# history:
|
||||
# 1995-09-01 fl Created
|
||||
# 1996-04-30 fl Added save
|
||||
# 1997-08-27 fl Fixed save of 1-bit images
|
||||
# 1998-03-06 fl Load P images as L where possible
|
||||
# 1998-07-03 fl Load P images as 1 where possible
|
||||
# 1998-12-29 fl Handle small palettes
|
||||
# 2002-12-30 fl Fixed load of 1-bit palette images
|
||||
# 2003-04-21 fl Fixed load of 1-bit monochrome images
|
||||
# 2003-04-23 fl Added limited support for BI_BITFIELDS compression
|
||||
#
|
||||
# Copyright (c) 1997-2003 by Secret Labs AB
|
||||
# Copyright (c) 1995-2003 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import os
|
||||
from typing import IO, Any
|
||||
|
||||
from . import Image, ImageFile, ImagePalette
|
||||
from ._binary import i16le as i16
|
||||
from ._binary import i32le as i32
|
||||
from ._binary import o8
|
||||
from ._binary import o16le as o16
|
||||
from ._binary import o32le as o32
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
# Read BMP file
|
||||
|
||||
BIT2MODE = {
|
||||
# bits => mode, rawmode
|
||||
1: ("P", "P;1"),
|
||||
4: ("P", "P;4"),
|
||||
8: ("P", "P"),
|
||||
16: ("RGB", "BGR;15"),
|
||||
24: ("RGB", "BGR"),
|
||||
32: ("RGB", "BGRX"),
|
||||
}
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:2] == b"BM"
|
||||
|
||||
|
||||
def _dib_accept(prefix: bytes) -> bool:
|
||||
return i32(prefix) in [12, 40, 52, 56, 64, 108, 124]
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# Image plugin for the Windows BMP format.
|
||||
# =============================================================================
|
||||
class BmpImageFile(ImageFile.ImageFile):
|
||||
"""Image plugin for the Windows Bitmap format (BMP)"""
|
||||
|
||||
# ------------------------------------------------------------- Description
|
||||
format_description = "Windows Bitmap"
|
||||
format = "BMP"
|
||||
|
||||
# -------------------------------------------------- BMP Compression values
|
||||
COMPRESSIONS = {"RAW": 0, "RLE8": 1, "RLE4": 2, "BITFIELDS": 3, "JPEG": 4, "PNG": 5}
|
||||
for k, v in COMPRESSIONS.items():
|
||||
vars()[k] = v
|
||||
|
||||
def _bitmap(self, header: int = 0, offset: int = 0) -> None:
|
||||
"""Read relevant info about the BMP"""
|
||||
read, seek = self.fp.read, self.fp.seek
|
||||
if header:
|
||||
seek(header)
|
||||
# read bmp header size @offset 14 (this is part of the header size)
|
||||
file_info: dict[str, bool | int | tuple[int, ...]] = {
|
||||
"header_size": i32(read(4)),
|
||||
"direction": -1,
|
||||
}
|
||||
|
||||
# -------------------- If requested, read header at a specific position
|
||||
# read the rest of the bmp header, without its size
|
||||
assert isinstance(file_info["header_size"], int)
|
||||
header_data = ImageFile._safe_read(self.fp, file_info["header_size"] - 4)
|
||||
|
||||
# ------------------------------- Windows Bitmap v2, IBM OS/2 Bitmap v1
|
||||
# ----- This format has different offsets because of width/height types
|
||||
# 12: BITMAPCOREHEADER/OS21XBITMAPHEADER
|
||||
if file_info["header_size"] == 12:
|
||||
file_info["width"] = i16(header_data, 0)
|
||||
file_info["height"] = i16(header_data, 2)
|
||||
file_info["planes"] = i16(header_data, 4)
|
||||
file_info["bits"] = i16(header_data, 6)
|
||||
file_info["compression"] = self.COMPRESSIONS["RAW"]
|
||||
file_info["palette_padding"] = 3
|
||||
|
||||
# --------------------------------------------- Windows Bitmap v3 to v5
|
||||
# 40: BITMAPINFOHEADER
|
||||
# 52: BITMAPV2HEADER
|
||||
# 56: BITMAPV3HEADER
|
||||
# 64: BITMAPCOREHEADER2/OS22XBITMAPHEADER
|
||||
# 108: BITMAPV4HEADER
|
||||
# 124: BITMAPV5HEADER
|
||||
elif file_info["header_size"] in (40, 52, 56, 64, 108, 124):
|
||||
file_info["y_flip"] = header_data[7] == 0xFF
|
||||
file_info["direction"] = 1 if file_info["y_flip"] else -1
|
||||
file_info["width"] = i32(header_data, 0)
|
||||
file_info["height"] = (
|
||||
i32(header_data, 4)
|
||||
if not file_info["y_flip"]
|
||||
else 2**32 - i32(header_data, 4)
|
||||
)
|
||||
file_info["planes"] = i16(header_data, 8)
|
||||
file_info["bits"] = i16(header_data, 10)
|
||||
file_info["compression"] = i32(header_data, 12)
|
||||
# byte size of pixel data
|
||||
file_info["data_size"] = i32(header_data, 16)
|
||||
file_info["pixels_per_meter"] = (
|
||||
i32(header_data, 20),
|
||||
i32(header_data, 24),
|
||||
)
|
||||
file_info["colors"] = i32(header_data, 28)
|
||||
file_info["palette_padding"] = 4
|
||||
assert isinstance(file_info["pixels_per_meter"], tuple)
|
||||
self.info["dpi"] = tuple(x / 39.3701 for x in file_info["pixels_per_meter"])
|
||||
if file_info["compression"] == self.COMPRESSIONS["BITFIELDS"]:
|
||||
masks = ["r_mask", "g_mask", "b_mask"]
|
||||
if len(header_data) >= 48:
|
||||
if len(header_data) >= 52:
|
||||
masks.append("a_mask")
|
||||
else:
|
||||
file_info["a_mask"] = 0x0
|
||||
for idx, mask in enumerate(masks):
|
||||
file_info[mask] = i32(header_data, 36 + idx * 4)
|
||||
else:
|
||||
# 40 byte headers only have the three components in the
|
||||
# bitfields masks, ref:
|
||||
# https://msdn.microsoft.com/en-us/library/windows/desktop/dd183376(v=vs.85).aspx
|
||||
# See also
|
||||
# https://github.com/python-pillow/Pillow/issues/1293
|
||||
# There is a 4th component in the RGBQuad, in the alpha
|
||||
# location, but it is listed as a reserved component,
|
||||
# and it is not generally an alpha channel
|
||||
file_info["a_mask"] = 0x0
|
||||
for mask in masks:
|
||||
file_info[mask] = i32(read(4))
|
||||
assert isinstance(file_info["r_mask"], int)
|
||||
assert isinstance(file_info["g_mask"], int)
|
||||
assert isinstance(file_info["b_mask"], int)
|
||||
assert isinstance(file_info["a_mask"], int)
|
||||
file_info["rgb_mask"] = (
|
||||
file_info["r_mask"],
|
||||
file_info["g_mask"],
|
||||
file_info["b_mask"],
|
||||
)
|
||||
file_info["rgba_mask"] = (
|
||||
file_info["r_mask"],
|
||||
file_info["g_mask"],
|
||||
file_info["b_mask"],
|
||||
file_info["a_mask"],
|
||||
)
|
||||
else:
|
||||
msg = f"Unsupported BMP header type ({file_info['header_size']})"
|
||||
raise OSError(msg)
|
||||
|
||||
# ------------------ Special case : header is reported 40, which
|
||||
# ---------------------- is shorter than real size for bpp >= 16
|
||||
assert isinstance(file_info["width"], int)
|
||||
assert isinstance(file_info["height"], int)
|
||||
self._size = file_info["width"], file_info["height"]
|
||||
|
||||
# ------- If color count was not found in the header, compute from bits
|
||||
assert isinstance(file_info["bits"], int)
|
||||
file_info["colors"] = (
|
||||
file_info["colors"]
|
||||
if file_info.get("colors", 0)
|
||||
else (1 << file_info["bits"])
|
||||
)
|
||||
assert isinstance(file_info["colors"], int)
|
||||
if offset == 14 + file_info["header_size"] and file_info["bits"] <= 8:
|
||||
offset += 4 * file_info["colors"]
|
||||
|
||||
# ---------------------- Check bit depth for unusual unsupported values
|
||||
self._mode, raw_mode = BIT2MODE.get(file_info["bits"], ("", ""))
|
||||
if not self.mode:
|
||||
msg = f"Unsupported BMP pixel depth ({file_info['bits']})"
|
||||
raise OSError(msg)
|
||||
|
||||
# ---------------- Process BMP with Bitfields compression (not palette)
|
||||
decoder_name = "raw"
|
||||
if file_info["compression"] == self.COMPRESSIONS["BITFIELDS"]:
|
||||
SUPPORTED: dict[int, list[tuple[int, ...]]] = {
|
||||
32: [
|
||||
(0xFF0000, 0xFF00, 0xFF, 0x0),
|
||||
(0xFF000000, 0xFF0000, 0xFF00, 0x0),
|
||||
(0xFF000000, 0xFF00, 0xFF, 0x0),
|
||||
(0xFF000000, 0xFF0000, 0xFF00, 0xFF),
|
||||
(0xFF, 0xFF00, 0xFF0000, 0xFF000000),
|
||||
(0xFF0000, 0xFF00, 0xFF, 0xFF000000),
|
||||
(0xFF000000, 0xFF00, 0xFF, 0xFF0000),
|
||||
(0x0, 0x0, 0x0, 0x0),
|
||||
],
|
||||
24: [(0xFF0000, 0xFF00, 0xFF)],
|
||||
16: [(0xF800, 0x7E0, 0x1F), (0x7C00, 0x3E0, 0x1F)],
|
||||
}
|
||||
MASK_MODES = {
|
||||
(32, (0xFF0000, 0xFF00, 0xFF, 0x0)): "BGRX",
|
||||
(32, (0xFF000000, 0xFF0000, 0xFF00, 0x0)): "XBGR",
|
||||
(32, (0xFF000000, 0xFF00, 0xFF, 0x0)): "BGXR",
|
||||
(32, (0xFF000000, 0xFF0000, 0xFF00, 0xFF)): "ABGR",
|
||||
(32, (0xFF, 0xFF00, 0xFF0000, 0xFF000000)): "RGBA",
|
||||
(32, (0xFF0000, 0xFF00, 0xFF, 0xFF000000)): "BGRA",
|
||||
(32, (0xFF000000, 0xFF00, 0xFF, 0xFF0000)): "BGAR",
|
||||
(32, (0x0, 0x0, 0x0, 0x0)): "BGRA",
|
||||
(24, (0xFF0000, 0xFF00, 0xFF)): "BGR",
|
||||
(16, (0xF800, 0x7E0, 0x1F)): "BGR;16",
|
||||
(16, (0x7C00, 0x3E0, 0x1F)): "BGR;15",
|
||||
}
|
||||
if file_info["bits"] in SUPPORTED:
|
||||
if (
|
||||
file_info["bits"] == 32
|
||||
and file_info["rgba_mask"] in SUPPORTED[file_info["bits"]]
|
||||
):
|
||||
assert isinstance(file_info["rgba_mask"], tuple)
|
||||
raw_mode = MASK_MODES[(file_info["bits"], file_info["rgba_mask"])]
|
||||
self._mode = "RGBA" if "A" in raw_mode else self.mode
|
||||
elif (
|
||||
file_info["bits"] in (24, 16)
|
||||
and file_info["rgb_mask"] in SUPPORTED[file_info["bits"]]
|
||||
):
|
||||
assert isinstance(file_info["rgb_mask"], tuple)
|
||||
raw_mode = MASK_MODES[(file_info["bits"], file_info["rgb_mask"])]
|
||||
else:
|
||||
msg = "Unsupported BMP bitfields layout"
|
||||
raise OSError(msg)
|
||||
else:
|
||||
msg = "Unsupported BMP bitfields layout"
|
||||
raise OSError(msg)
|
||||
elif file_info["compression"] == self.COMPRESSIONS["RAW"]:
|
||||
if file_info["bits"] == 32 and header == 22: # 32-bit .cur offset
|
||||
raw_mode, self._mode = "BGRA", "RGBA"
|
||||
elif file_info["compression"] in (
|
||||
self.COMPRESSIONS["RLE8"],
|
||||
self.COMPRESSIONS["RLE4"],
|
||||
):
|
||||
decoder_name = "bmp_rle"
|
||||
else:
|
||||
msg = f"Unsupported BMP compression ({file_info['compression']})"
|
||||
raise OSError(msg)
|
||||
|
||||
# --------------- Once the header is processed, process the palette/LUT
|
||||
if self.mode == "P": # Paletted for 1, 4 and 8 bit images
|
||||
# ---------------------------------------------------- 1-bit images
|
||||
if not (0 < file_info["colors"] <= 65536):
|
||||
msg = f"Unsupported BMP Palette size ({file_info['colors']})"
|
||||
raise OSError(msg)
|
||||
else:
|
||||
assert isinstance(file_info["palette_padding"], int)
|
||||
padding = file_info["palette_padding"]
|
||||
palette = read(padding * file_info["colors"])
|
||||
grayscale = True
|
||||
indices = (
|
||||
(0, 255)
|
||||
if file_info["colors"] == 2
|
||||
else list(range(file_info["colors"]))
|
||||
)
|
||||
|
||||
# ----------------- Check if grayscale and ignore palette if so
|
||||
for ind, val in enumerate(indices):
|
||||
rgb = palette[ind * padding : ind * padding + 3]
|
||||
if rgb != o8(val) * 3:
|
||||
grayscale = False
|
||||
|
||||
# ------- If all colors are gray, white or black, ditch palette
|
||||
if grayscale:
|
||||
self._mode = "1" if file_info["colors"] == 2 else "L"
|
||||
raw_mode = self.mode
|
||||
else:
|
||||
self._mode = "P"
|
||||
self.palette = ImagePalette.raw(
|
||||
"BGRX" if padding == 4 else "BGR", palette
|
||||
)
|
||||
|
||||
# ---------------------------- Finally set the tile data for the plugin
|
||||
self.info["compression"] = file_info["compression"]
|
||||
args: list[Any] = [raw_mode]
|
||||
if decoder_name == "bmp_rle":
|
||||
args.append(file_info["compression"] == self.COMPRESSIONS["RLE4"])
|
||||
else:
|
||||
assert isinstance(file_info["width"], int)
|
||||
args.append(((file_info["width"] * file_info["bits"] + 31) >> 3) & (~3))
|
||||
args.append(file_info["direction"])
|
||||
self.tile = [
|
||||
ImageFile._Tile(
|
||||
decoder_name,
|
||||
(0, 0, file_info["width"], file_info["height"]),
|
||||
offset or self.fp.tell(),
|
||||
tuple(args),
|
||||
)
|
||||
]
|
||||
|
||||
def _open(self) -> None:
|
||||
"""Open file, check magic number and read header"""
|
||||
# read 14 bytes: magic number, filesize, reserved, header final offset
|
||||
head_data = self.fp.read(14)
|
||||
# choke if the file does not have the required magic bytes
|
||||
if not _accept(head_data):
|
||||
msg = "Not a BMP file"
|
||||
raise SyntaxError(msg)
|
||||
# read the start position of the BMP image data (u32)
|
||||
offset = i32(head_data, 10)
|
||||
# load bitmap information (offset=raster info)
|
||||
self._bitmap(offset=offset)
|
||||
|
||||
|
||||
class BmpRleDecoder(ImageFile.PyDecoder):
|
||||
_pulls_fd = True
|
||||
|
||||
def decode(self, buffer: bytes | Image.SupportsArrayInterface) -> tuple[int, int]:
|
||||
assert self.fd is not None
|
||||
rle4 = self.args[1]
|
||||
data = bytearray()
|
||||
x = 0
|
||||
dest_length = self.state.xsize * self.state.ysize
|
||||
while len(data) < dest_length:
|
||||
pixels = self.fd.read(1)
|
||||
byte = self.fd.read(1)
|
||||
if not pixels or not byte:
|
||||
break
|
||||
num_pixels = pixels[0]
|
||||
if num_pixels:
|
||||
# encoded mode
|
||||
if x + num_pixels > self.state.xsize:
|
||||
# Too much data for row
|
||||
num_pixels = max(0, self.state.xsize - x)
|
||||
if rle4:
|
||||
first_pixel = o8(byte[0] >> 4)
|
||||
second_pixel = o8(byte[0] & 0x0F)
|
||||
for index in range(num_pixels):
|
||||
if index % 2 == 0:
|
||||
data += first_pixel
|
||||
else:
|
||||
data += second_pixel
|
||||
else:
|
||||
data += byte * num_pixels
|
||||
x += num_pixels
|
||||
else:
|
||||
if byte[0] == 0:
|
||||
# end of line
|
||||
while len(data) % self.state.xsize != 0:
|
||||
data += b"\x00"
|
||||
x = 0
|
||||
elif byte[0] == 1:
|
||||
# end of bitmap
|
||||
break
|
||||
elif byte[0] == 2:
|
||||
# delta
|
||||
bytes_read = self.fd.read(2)
|
||||
if len(bytes_read) < 2:
|
||||
break
|
||||
right, up = self.fd.read(2)
|
||||
data += b"\x00" * (right + up * self.state.xsize)
|
||||
x = len(data) % self.state.xsize
|
||||
else:
|
||||
# absolute mode
|
||||
if rle4:
|
||||
# 2 pixels per byte
|
||||
byte_count = byte[0] // 2
|
||||
bytes_read = self.fd.read(byte_count)
|
||||
for byte_read in bytes_read:
|
||||
data += o8(byte_read >> 4)
|
||||
data += o8(byte_read & 0x0F)
|
||||
else:
|
||||
byte_count = byte[0]
|
||||
bytes_read = self.fd.read(byte_count)
|
||||
data += bytes_read
|
||||
if len(bytes_read) < byte_count:
|
||||
break
|
||||
x += byte[0]
|
||||
|
||||
# align to 16-bit word boundary
|
||||
if self.fd.tell() % 2 != 0:
|
||||
self.fd.seek(1, os.SEEK_CUR)
|
||||
rawmode = "L" if self.mode == "L" else "P"
|
||||
self.set_as_raw(bytes(data), rawmode, (0, self.args[-1]))
|
||||
return -1, 0
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# Image plugin for the DIB format (BMP alias)
|
||||
# =============================================================================
|
||||
class DibImageFile(BmpImageFile):
|
||||
format = "DIB"
|
||||
format_description = "Windows Bitmap"
|
||||
|
||||
def _open(self) -> None:
|
||||
self._bitmap()
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
# Write BMP file
|
||||
|
||||
|
||||
SAVE = {
|
||||
"1": ("1", 1, 2),
|
||||
"L": ("L", 8, 256),
|
||||
"P": ("P", 8, 256),
|
||||
"RGB": ("BGR", 24, 0),
|
||||
"RGBA": ("BGRA", 32, 0),
|
||||
}
|
||||
|
||||
|
||||
def _dib_save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
_save(im, fp, filename, False)
|
||||
|
||||
|
||||
def _save(
|
||||
im: Image.Image, fp: IO[bytes], filename: str | bytes, bitmap_header: bool = True
|
||||
) -> None:
|
||||
try:
|
||||
rawmode, bits, colors = SAVE[im.mode]
|
||||
except KeyError as e:
|
||||
msg = f"cannot write mode {im.mode} as BMP"
|
||||
raise OSError(msg) from e
|
||||
|
||||
info = im.encoderinfo
|
||||
|
||||
dpi = info.get("dpi", (96, 96))
|
||||
|
||||
# 1 meter == 39.3701 inches
|
||||
ppm = tuple(int(x * 39.3701 + 0.5) for x in dpi)
|
||||
|
||||
stride = ((im.size[0] * bits + 7) // 8 + 3) & (~3)
|
||||
header = 40 # or 64 for OS/2 version 2
|
||||
image = stride * im.size[1]
|
||||
|
||||
if im.mode == "1":
|
||||
palette = b"".join(o8(i) * 4 for i in (0, 255))
|
||||
elif im.mode == "L":
|
||||
palette = b"".join(o8(i) * 4 for i in range(256))
|
||||
elif im.mode == "P":
|
||||
palette = im.im.getpalette("RGB", "BGRX")
|
||||
colors = len(palette) // 4
|
||||
else:
|
||||
palette = None
|
||||
|
||||
# bitmap header
|
||||
if bitmap_header:
|
||||
offset = 14 + header + colors * 4
|
||||
file_size = offset + image
|
||||
if file_size > 2**32 - 1:
|
||||
msg = "File size is too large for the BMP format"
|
||||
raise ValueError(msg)
|
||||
fp.write(
|
||||
b"BM" # file type (magic)
|
||||
+ o32(file_size) # file size
|
||||
+ o32(0) # reserved
|
||||
+ o32(offset) # image data offset
|
||||
)
|
||||
|
||||
# bitmap info header
|
||||
fp.write(
|
||||
o32(header) # info header size
|
||||
+ o32(im.size[0]) # width
|
||||
+ o32(im.size[1]) # height
|
||||
+ o16(1) # planes
|
||||
+ o16(bits) # depth
|
||||
+ o32(0) # compression (0=uncompressed)
|
||||
+ o32(image) # size of bitmap
|
||||
+ o32(ppm[0]) # resolution
|
||||
+ o32(ppm[1]) # resolution
|
||||
+ o32(colors) # colors used
|
||||
+ o32(colors) # colors important
|
||||
)
|
||||
|
||||
fp.write(b"\0" * (header - 40)) # padding (for OS/2 format)
|
||||
|
||||
if palette:
|
||||
fp.write(palette)
|
||||
|
||||
ImageFile._save(
|
||||
im, fp, [ImageFile._Tile("raw", (0, 0) + im.size, 0, (rawmode, stride, -1))]
|
||||
)
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
# Registry
|
||||
|
||||
|
||||
Image.register_open(BmpImageFile.format, BmpImageFile, _accept)
|
||||
Image.register_save(BmpImageFile.format, _save)
|
||||
|
||||
Image.register_extension(BmpImageFile.format, ".bmp")
|
||||
|
||||
Image.register_mime(BmpImageFile.format, "image/bmp")
|
||||
|
||||
Image.register_decoder("bmp_rle", BmpRleDecoder)
|
||||
|
||||
Image.register_open(DibImageFile.format, DibImageFile, _dib_accept)
|
||||
Image.register_save(DibImageFile.format, _dib_save)
|
||||
|
||||
Image.register_extension(DibImageFile.format, ".dib")
|
||||
|
||||
Image.register_mime(DibImageFile.format, "image/bmp")
|
||||
@ -0,0 +1,76 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# BUFR stub adapter
|
||||
#
|
||||
# Copyright (c) 1996-2003 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import IO
|
||||
|
||||
from . import Image, ImageFile
|
||||
|
||||
_handler = None
|
||||
|
||||
|
||||
def register_handler(handler: ImageFile.StubHandler | None) -> None:
|
||||
"""
|
||||
Install application-specific BUFR image handler.
|
||||
|
||||
:param handler: Handler object.
|
||||
"""
|
||||
global _handler
|
||||
_handler = handler
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Image adapter
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == b"BUFR" or prefix[:4] == b"ZCZC"
|
||||
|
||||
|
||||
class BufrStubImageFile(ImageFile.StubImageFile):
|
||||
format = "BUFR"
|
||||
format_description = "BUFR"
|
||||
|
||||
def _open(self) -> None:
|
||||
offset = self.fp.tell()
|
||||
|
||||
if not _accept(self.fp.read(4)):
|
||||
msg = "Not a BUFR file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
self.fp.seek(offset)
|
||||
|
||||
# make something up
|
||||
self._mode = "F"
|
||||
self._size = 1, 1
|
||||
|
||||
loader = self._load()
|
||||
if loader:
|
||||
loader.open(self)
|
||||
|
||||
def _load(self) -> ImageFile.StubHandler | None:
|
||||
return _handler
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
if _handler is None or not hasattr(_handler, "save"):
|
||||
msg = "BUFR save handler not installed"
|
||||
raise OSError(msg)
|
||||
_handler.save(im, fp, filename)
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Registry
|
||||
|
||||
Image.register_open(BufrStubImageFile.format, BufrStubImageFile, _accept)
|
||||
Image.register_save(BufrStubImageFile.format, _save)
|
||||
|
||||
Image.register_extension(BufrStubImageFile.format, ".bufr")
|
||||
173
testclass/lib/python3.12/site-packages/PIL/ContainerIO.py
Normal file
173
testclass/lib/python3.12/site-packages/PIL/ContainerIO.py
Normal file
@ -0,0 +1,173 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# a class to read from a container file
|
||||
#
|
||||
# History:
|
||||
# 1995-06-18 fl Created
|
||||
# 1995-09-07 fl Added readline(), readlines()
|
||||
#
|
||||
# Copyright (c) 1997-2001 by Secret Labs AB
|
||||
# Copyright (c) 1995 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import io
|
||||
from collections.abc import Iterable
|
||||
from typing import IO, AnyStr, NoReturn
|
||||
|
||||
|
||||
class ContainerIO(IO[AnyStr]):
|
||||
"""
|
||||
A file object that provides read access to a part of an existing
|
||||
file (for example a TAR file).
|
||||
"""
|
||||
|
||||
def __init__(self, file: IO[AnyStr], offset: int, length: int) -> None:
|
||||
"""
|
||||
Create file object.
|
||||
|
||||
:param file: Existing file.
|
||||
:param offset: Start of region, in bytes.
|
||||
:param length: Size of region, in bytes.
|
||||
"""
|
||||
self.fh: IO[AnyStr] = file
|
||||
self.pos = 0
|
||||
self.offset = offset
|
||||
self.length = length
|
||||
self.fh.seek(offset)
|
||||
|
||||
##
|
||||
# Always false.
|
||||
|
||||
def isatty(self) -> bool:
|
||||
return False
|
||||
|
||||
def seekable(self) -> bool:
|
||||
return True
|
||||
|
||||
def seek(self, offset: int, mode: int = io.SEEK_SET) -> int:
|
||||
"""
|
||||
Move file pointer.
|
||||
|
||||
:param offset: Offset in bytes.
|
||||
:param mode: Starting position. Use 0 for beginning of region, 1
|
||||
for current offset, and 2 for end of region. You cannot move
|
||||
the pointer outside the defined region.
|
||||
:returns: Offset from start of region, in bytes.
|
||||
"""
|
||||
if mode == 1:
|
||||
self.pos = self.pos + offset
|
||||
elif mode == 2:
|
||||
self.pos = self.length + offset
|
||||
else:
|
||||
self.pos = offset
|
||||
# clamp
|
||||
self.pos = max(0, min(self.pos, self.length))
|
||||
self.fh.seek(self.offset + self.pos)
|
||||
return self.pos
|
||||
|
||||
def tell(self) -> int:
|
||||
"""
|
||||
Get current file pointer.
|
||||
|
||||
:returns: Offset from start of region, in bytes.
|
||||
"""
|
||||
return self.pos
|
||||
|
||||
def readable(self) -> bool:
|
||||
return True
|
||||
|
||||
def read(self, n: int = -1) -> AnyStr:
|
||||
"""
|
||||
Read data.
|
||||
|
||||
:param n: Number of bytes to read. If omitted, zero or negative,
|
||||
read until end of region.
|
||||
:returns: An 8-bit string.
|
||||
"""
|
||||
if n > 0:
|
||||
n = min(n, self.length - self.pos)
|
||||
else:
|
||||
n = self.length - self.pos
|
||||
if n <= 0: # EOF
|
||||
return b"" if "b" in self.fh.mode else "" # type: ignore[return-value]
|
||||
self.pos = self.pos + n
|
||||
return self.fh.read(n)
|
||||
|
||||
def readline(self, n: int = -1) -> AnyStr:
|
||||
"""
|
||||
Read a line of text.
|
||||
|
||||
:param n: Number of bytes to read. If omitted, zero or negative,
|
||||
read until end of line.
|
||||
:returns: An 8-bit string.
|
||||
"""
|
||||
s: AnyStr = b"" if "b" in self.fh.mode else "" # type: ignore[assignment]
|
||||
newline_character = b"\n" if "b" in self.fh.mode else "\n"
|
||||
while True:
|
||||
c = self.read(1)
|
||||
if not c:
|
||||
break
|
||||
s = s + c
|
||||
if c == newline_character or len(s) == n:
|
||||
break
|
||||
return s
|
||||
|
||||
def readlines(self, n: int | None = -1) -> list[AnyStr]:
|
||||
"""
|
||||
Read multiple lines of text.
|
||||
|
||||
:param n: Number of lines to read. If omitted, zero, negative or None,
|
||||
read until end of region.
|
||||
:returns: A list of 8-bit strings.
|
||||
"""
|
||||
lines = []
|
||||
while True:
|
||||
s = self.readline()
|
||||
if not s:
|
||||
break
|
||||
lines.append(s)
|
||||
if len(lines) == n:
|
||||
break
|
||||
return lines
|
||||
|
||||
def writable(self) -> bool:
|
||||
return False
|
||||
|
||||
def write(self, b: AnyStr) -> NoReturn:
|
||||
raise NotImplementedError()
|
||||
|
||||
def writelines(self, lines: Iterable[AnyStr]) -> NoReturn:
|
||||
raise NotImplementedError()
|
||||
|
||||
def truncate(self, size: int | None = None) -> int:
|
||||
raise NotImplementedError()
|
||||
|
||||
def __enter__(self) -> ContainerIO[AnyStr]:
|
||||
return self
|
||||
|
||||
def __exit__(self, *args: object) -> None:
|
||||
self.close()
|
||||
|
||||
def __iter__(self) -> ContainerIO[AnyStr]:
|
||||
return self
|
||||
|
||||
def __next__(self) -> AnyStr:
|
||||
line = self.readline()
|
||||
if not line:
|
||||
msg = "end of region"
|
||||
raise StopIteration(msg)
|
||||
return line
|
||||
|
||||
def fileno(self) -> int:
|
||||
return self.fh.fileno()
|
||||
|
||||
def flush(self) -> None:
|
||||
self.fh.flush()
|
||||
|
||||
def close(self) -> None:
|
||||
self.fh.close()
|
||||
75
testclass/lib/python3.12/site-packages/PIL/CurImagePlugin.py
Normal file
75
testclass/lib/python3.12/site-packages/PIL/CurImagePlugin.py
Normal file
@ -0,0 +1,75 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# Windows Cursor support for PIL
|
||||
#
|
||||
# notes:
|
||||
# uses BmpImagePlugin.py to read the bitmap data.
|
||||
#
|
||||
# history:
|
||||
# 96-05-27 fl Created
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997.
|
||||
# Copyright (c) Fredrik Lundh 1996.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
from . import BmpImagePlugin, Image, ImageFile
|
||||
from ._binary import i16le as i16
|
||||
from ._binary import i32le as i32
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == b"\0\0\2\0"
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for Windows Cursor files.
|
||||
|
||||
|
||||
class CurImageFile(BmpImagePlugin.BmpImageFile):
|
||||
format = "CUR"
|
||||
format_description = "Windows Cursor"
|
||||
|
||||
def _open(self) -> None:
|
||||
offset = self.fp.tell()
|
||||
|
||||
# check magic
|
||||
s = self.fp.read(6)
|
||||
if not _accept(s):
|
||||
msg = "not a CUR file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
# pick the largest cursor in the file
|
||||
m = b""
|
||||
for i in range(i16(s, 4)):
|
||||
s = self.fp.read(16)
|
||||
if not m:
|
||||
m = s
|
||||
elif s[0] > m[0] and s[1] > m[1]:
|
||||
m = s
|
||||
if not m:
|
||||
msg = "No cursors were found"
|
||||
raise TypeError(msg)
|
||||
|
||||
# load as bitmap
|
||||
self._bitmap(i32(m, 12) + offset)
|
||||
|
||||
# patch up the bitmap height
|
||||
self._size = self.size[0], self.size[1] // 2
|
||||
d, e, o, a = self.tile[0]
|
||||
self.tile[0] = ImageFile._Tile(d, (0, 0) + self.size, o, a)
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
Image.register_open(CurImageFile.format, CurImageFile, _accept)
|
||||
|
||||
Image.register_extension(CurImageFile.format, ".cur")
|
||||
80
testclass/lib/python3.12/site-packages/PIL/DcxImagePlugin.py
Normal file
80
testclass/lib/python3.12/site-packages/PIL/DcxImagePlugin.py
Normal file
@ -0,0 +1,80 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# DCX file handling
|
||||
#
|
||||
# DCX is a container file format defined by Intel, commonly used
|
||||
# for fax applications. Each DCX file consists of a directory
|
||||
# (a list of file offsets) followed by a set of (usually 1-bit)
|
||||
# PCX files.
|
||||
#
|
||||
# History:
|
||||
# 1995-09-09 fl Created
|
||||
# 1996-03-20 fl Properly derived from PcxImageFile.
|
||||
# 1998-07-15 fl Renamed offset attribute to avoid name clash
|
||||
# 2002-07-30 fl Fixed file handling
|
||||
#
|
||||
# Copyright (c) 1997-98 by Secret Labs AB.
|
||||
# Copyright (c) 1995-96 by Fredrik Lundh.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
from . import Image
|
||||
from ._binary import i32le as i32
|
||||
from .PcxImagePlugin import PcxImageFile
|
||||
|
||||
MAGIC = 0x3ADE68B1 # QUIZ: what's this value, then?
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return len(prefix) >= 4 and i32(prefix) == MAGIC
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for the Intel DCX format.
|
||||
|
||||
|
||||
class DcxImageFile(PcxImageFile):
|
||||
format = "DCX"
|
||||
format_description = "Intel DCX"
|
||||
_close_exclusive_fp_after_loading = False
|
||||
|
||||
def _open(self) -> None:
|
||||
# Header
|
||||
s = self.fp.read(4)
|
||||
if not _accept(s):
|
||||
msg = "not a DCX file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
# Component directory
|
||||
self._offset = []
|
||||
for i in range(1024):
|
||||
offset = i32(self.fp.read(4))
|
||||
if not offset:
|
||||
break
|
||||
self._offset.append(offset)
|
||||
|
||||
self._fp = self.fp
|
||||
self.frame = -1
|
||||
self.n_frames = len(self._offset)
|
||||
self.is_animated = self.n_frames > 1
|
||||
self.seek(0)
|
||||
|
||||
def seek(self, frame: int) -> None:
|
||||
if not self._seek_check(frame):
|
||||
return
|
||||
self.frame = frame
|
||||
self.fp = self._fp
|
||||
self.fp.seek(self._offset[frame])
|
||||
PcxImageFile._open(self)
|
||||
|
||||
def tell(self) -> int:
|
||||
return self.frame
|
||||
|
||||
|
||||
Image.register_open(DcxImageFile.format, DcxImageFile, _accept)
|
||||
|
||||
Image.register_extension(DcxImageFile.format, ".dcx")
|
||||
573
testclass/lib/python3.12/site-packages/PIL/DdsImagePlugin.py
Normal file
573
testclass/lib/python3.12/site-packages/PIL/DdsImagePlugin.py
Normal file
@ -0,0 +1,573 @@
|
||||
"""
|
||||
A Pillow loader for .dds files (S3TC-compressed aka DXTC)
|
||||
Jerome Leclanche <jerome@leclan.ch>
|
||||
|
||||
Documentation:
|
||||
https://web.archive.org/web/20170802060935/http://oss.sgi.com/projects/ogl-sample/registry/EXT/texture_compression_s3tc.txt
|
||||
|
||||
The contents of this file are hereby released in the public domain (CC0)
|
||||
Full text of the CC0 license:
|
||||
https://creativecommons.org/publicdomain/zero/1.0/
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import io
|
||||
import struct
|
||||
import sys
|
||||
from enum import IntEnum, IntFlag
|
||||
from typing import IO
|
||||
|
||||
from . import Image, ImageFile, ImagePalette
|
||||
from ._binary import i32le as i32
|
||||
from ._binary import o8
|
||||
from ._binary import o32le as o32
|
||||
|
||||
# Magic ("DDS ")
|
||||
DDS_MAGIC = 0x20534444
|
||||
|
||||
|
||||
# DDS flags
|
||||
class DDSD(IntFlag):
|
||||
CAPS = 0x1
|
||||
HEIGHT = 0x2
|
||||
WIDTH = 0x4
|
||||
PITCH = 0x8
|
||||
PIXELFORMAT = 0x1000
|
||||
MIPMAPCOUNT = 0x20000
|
||||
LINEARSIZE = 0x80000
|
||||
DEPTH = 0x800000
|
||||
|
||||
|
||||
# DDS caps
|
||||
class DDSCAPS(IntFlag):
|
||||
COMPLEX = 0x8
|
||||
TEXTURE = 0x1000
|
||||
MIPMAP = 0x400000
|
||||
|
||||
|
||||
class DDSCAPS2(IntFlag):
|
||||
CUBEMAP = 0x200
|
||||
CUBEMAP_POSITIVEX = 0x400
|
||||
CUBEMAP_NEGATIVEX = 0x800
|
||||
CUBEMAP_POSITIVEY = 0x1000
|
||||
CUBEMAP_NEGATIVEY = 0x2000
|
||||
CUBEMAP_POSITIVEZ = 0x4000
|
||||
CUBEMAP_NEGATIVEZ = 0x8000
|
||||
VOLUME = 0x200000
|
||||
|
||||
|
||||
# Pixel Format
|
||||
class DDPF(IntFlag):
|
||||
ALPHAPIXELS = 0x1
|
||||
ALPHA = 0x2
|
||||
FOURCC = 0x4
|
||||
PALETTEINDEXED8 = 0x20
|
||||
RGB = 0x40
|
||||
LUMINANCE = 0x20000
|
||||
|
||||
|
||||
# dxgiformat.h
|
||||
class DXGI_FORMAT(IntEnum):
|
||||
UNKNOWN = 0
|
||||
R32G32B32A32_TYPELESS = 1
|
||||
R32G32B32A32_FLOAT = 2
|
||||
R32G32B32A32_UINT = 3
|
||||
R32G32B32A32_SINT = 4
|
||||
R32G32B32_TYPELESS = 5
|
||||
R32G32B32_FLOAT = 6
|
||||
R32G32B32_UINT = 7
|
||||
R32G32B32_SINT = 8
|
||||
R16G16B16A16_TYPELESS = 9
|
||||
R16G16B16A16_FLOAT = 10
|
||||
R16G16B16A16_UNORM = 11
|
||||
R16G16B16A16_UINT = 12
|
||||
R16G16B16A16_SNORM = 13
|
||||
R16G16B16A16_SINT = 14
|
||||
R32G32_TYPELESS = 15
|
||||
R32G32_FLOAT = 16
|
||||
R32G32_UINT = 17
|
||||
R32G32_SINT = 18
|
||||
R32G8X24_TYPELESS = 19
|
||||
D32_FLOAT_S8X24_UINT = 20
|
||||
R32_FLOAT_X8X24_TYPELESS = 21
|
||||
X32_TYPELESS_G8X24_UINT = 22
|
||||
R10G10B10A2_TYPELESS = 23
|
||||
R10G10B10A2_UNORM = 24
|
||||
R10G10B10A2_UINT = 25
|
||||
R11G11B10_FLOAT = 26
|
||||
R8G8B8A8_TYPELESS = 27
|
||||
R8G8B8A8_UNORM = 28
|
||||
R8G8B8A8_UNORM_SRGB = 29
|
||||
R8G8B8A8_UINT = 30
|
||||
R8G8B8A8_SNORM = 31
|
||||
R8G8B8A8_SINT = 32
|
||||
R16G16_TYPELESS = 33
|
||||
R16G16_FLOAT = 34
|
||||
R16G16_UNORM = 35
|
||||
R16G16_UINT = 36
|
||||
R16G16_SNORM = 37
|
||||
R16G16_SINT = 38
|
||||
R32_TYPELESS = 39
|
||||
D32_FLOAT = 40
|
||||
R32_FLOAT = 41
|
||||
R32_UINT = 42
|
||||
R32_SINT = 43
|
||||
R24G8_TYPELESS = 44
|
||||
D24_UNORM_S8_UINT = 45
|
||||
R24_UNORM_X8_TYPELESS = 46
|
||||
X24_TYPELESS_G8_UINT = 47
|
||||
R8G8_TYPELESS = 48
|
||||
R8G8_UNORM = 49
|
||||
R8G8_UINT = 50
|
||||
R8G8_SNORM = 51
|
||||
R8G8_SINT = 52
|
||||
R16_TYPELESS = 53
|
||||
R16_FLOAT = 54
|
||||
D16_UNORM = 55
|
||||
R16_UNORM = 56
|
||||
R16_UINT = 57
|
||||
R16_SNORM = 58
|
||||
R16_SINT = 59
|
||||
R8_TYPELESS = 60
|
||||
R8_UNORM = 61
|
||||
R8_UINT = 62
|
||||
R8_SNORM = 63
|
||||
R8_SINT = 64
|
||||
A8_UNORM = 65
|
||||
R1_UNORM = 66
|
||||
R9G9B9E5_SHAREDEXP = 67
|
||||
R8G8_B8G8_UNORM = 68
|
||||
G8R8_G8B8_UNORM = 69
|
||||
BC1_TYPELESS = 70
|
||||
BC1_UNORM = 71
|
||||
BC1_UNORM_SRGB = 72
|
||||
BC2_TYPELESS = 73
|
||||
BC2_UNORM = 74
|
||||
BC2_UNORM_SRGB = 75
|
||||
BC3_TYPELESS = 76
|
||||
BC3_UNORM = 77
|
||||
BC3_UNORM_SRGB = 78
|
||||
BC4_TYPELESS = 79
|
||||
BC4_UNORM = 80
|
||||
BC4_SNORM = 81
|
||||
BC5_TYPELESS = 82
|
||||
BC5_UNORM = 83
|
||||
BC5_SNORM = 84
|
||||
B5G6R5_UNORM = 85
|
||||
B5G5R5A1_UNORM = 86
|
||||
B8G8R8A8_UNORM = 87
|
||||
B8G8R8X8_UNORM = 88
|
||||
R10G10B10_XR_BIAS_A2_UNORM = 89
|
||||
B8G8R8A8_TYPELESS = 90
|
||||
B8G8R8A8_UNORM_SRGB = 91
|
||||
B8G8R8X8_TYPELESS = 92
|
||||
B8G8R8X8_UNORM_SRGB = 93
|
||||
BC6H_TYPELESS = 94
|
||||
BC6H_UF16 = 95
|
||||
BC6H_SF16 = 96
|
||||
BC7_TYPELESS = 97
|
||||
BC7_UNORM = 98
|
||||
BC7_UNORM_SRGB = 99
|
||||
AYUV = 100
|
||||
Y410 = 101
|
||||
Y416 = 102
|
||||
NV12 = 103
|
||||
P010 = 104
|
||||
P016 = 105
|
||||
OPAQUE_420 = 106
|
||||
YUY2 = 107
|
||||
Y210 = 108
|
||||
Y216 = 109
|
||||
NV11 = 110
|
||||
AI44 = 111
|
||||
IA44 = 112
|
||||
P8 = 113
|
||||
A8P8 = 114
|
||||
B4G4R4A4_UNORM = 115
|
||||
P208 = 130
|
||||
V208 = 131
|
||||
V408 = 132
|
||||
SAMPLER_FEEDBACK_MIN_MIP_OPAQUE = 189
|
||||
SAMPLER_FEEDBACK_MIP_REGION_USED_OPAQUE = 190
|
||||
|
||||
|
||||
class D3DFMT(IntEnum):
|
||||
UNKNOWN = 0
|
||||
R8G8B8 = 20
|
||||
A8R8G8B8 = 21
|
||||
X8R8G8B8 = 22
|
||||
R5G6B5 = 23
|
||||
X1R5G5B5 = 24
|
||||
A1R5G5B5 = 25
|
||||
A4R4G4B4 = 26
|
||||
R3G3B2 = 27
|
||||
A8 = 28
|
||||
A8R3G3B2 = 29
|
||||
X4R4G4B4 = 30
|
||||
A2B10G10R10 = 31
|
||||
A8B8G8R8 = 32
|
||||
X8B8G8R8 = 33
|
||||
G16R16 = 34
|
||||
A2R10G10B10 = 35
|
||||
A16B16G16R16 = 36
|
||||
A8P8 = 40
|
||||
P8 = 41
|
||||
L8 = 50
|
||||
A8L8 = 51
|
||||
A4L4 = 52
|
||||
V8U8 = 60
|
||||
L6V5U5 = 61
|
||||
X8L8V8U8 = 62
|
||||
Q8W8V8U8 = 63
|
||||
V16U16 = 64
|
||||
A2W10V10U10 = 67
|
||||
D16_LOCKABLE = 70
|
||||
D32 = 71
|
||||
D15S1 = 73
|
||||
D24S8 = 75
|
||||
D24X8 = 77
|
||||
D24X4S4 = 79
|
||||
D16 = 80
|
||||
D32F_LOCKABLE = 82
|
||||
D24FS8 = 83
|
||||
D32_LOCKABLE = 84
|
||||
S8_LOCKABLE = 85
|
||||
L16 = 81
|
||||
VERTEXDATA = 100
|
||||
INDEX16 = 101
|
||||
INDEX32 = 102
|
||||
Q16W16V16U16 = 110
|
||||
R16F = 111
|
||||
G16R16F = 112
|
||||
A16B16G16R16F = 113
|
||||
R32F = 114
|
||||
G32R32F = 115
|
||||
A32B32G32R32F = 116
|
||||
CxV8U8 = 117
|
||||
A1 = 118
|
||||
A2B10G10R10_XR_BIAS = 119
|
||||
BINARYBUFFER = 199
|
||||
|
||||
UYVY = i32(b"UYVY")
|
||||
R8G8_B8G8 = i32(b"RGBG")
|
||||
YUY2 = i32(b"YUY2")
|
||||
G8R8_G8B8 = i32(b"GRGB")
|
||||
DXT1 = i32(b"DXT1")
|
||||
DXT2 = i32(b"DXT2")
|
||||
DXT3 = i32(b"DXT3")
|
||||
DXT4 = i32(b"DXT4")
|
||||
DXT5 = i32(b"DXT5")
|
||||
DX10 = i32(b"DX10")
|
||||
BC4S = i32(b"BC4S")
|
||||
BC4U = i32(b"BC4U")
|
||||
BC5S = i32(b"BC5S")
|
||||
BC5U = i32(b"BC5U")
|
||||
ATI1 = i32(b"ATI1")
|
||||
ATI2 = i32(b"ATI2")
|
||||
MULTI2_ARGB8 = i32(b"MET1")
|
||||
|
||||
|
||||
# Backward compatibility layer
|
||||
module = sys.modules[__name__]
|
||||
for item in DDSD:
|
||||
assert item.name is not None
|
||||
setattr(module, f"DDSD_{item.name}", item.value)
|
||||
for item1 in DDSCAPS:
|
||||
assert item1.name is not None
|
||||
setattr(module, f"DDSCAPS_{item1.name}", item1.value)
|
||||
for item2 in DDSCAPS2:
|
||||
assert item2.name is not None
|
||||
setattr(module, f"DDSCAPS2_{item2.name}", item2.value)
|
||||
for item3 in DDPF:
|
||||
assert item3.name is not None
|
||||
setattr(module, f"DDPF_{item3.name}", item3.value)
|
||||
|
||||
DDS_FOURCC = DDPF.FOURCC
|
||||
DDS_RGB = DDPF.RGB
|
||||
DDS_RGBA = DDPF.RGB | DDPF.ALPHAPIXELS
|
||||
DDS_LUMINANCE = DDPF.LUMINANCE
|
||||
DDS_LUMINANCEA = DDPF.LUMINANCE | DDPF.ALPHAPIXELS
|
||||
DDS_ALPHA = DDPF.ALPHA
|
||||
DDS_PAL8 = DDPF.PALETTEINDEXED8
|
||||
|
||||
DDS_HEADER_FLAGS_TEXTURE = DDSD.CAPS | DDSD.HEIGHT | DDSD.WIDTH | DDSD.PIXELFORMAT
|
||||
DDS_HEADER_FLAGS_MIPMAP = DDSD.MIPMAPCOUNT
|
||||
DDS_HEADER_FLAGS_VOLUME = DDSD.DEPTH
|
||||
DDS_HEADER_FLAGS_PITCH = DDSD.PITCH
|
||||
DDS_HEADER_FLAGS_LINEARSIZE = DDSD.LINEARSIZE
|
||||
|
||||
DDS_HEIGHT = DDSD.HEIGHT
|
||||
DDS_WIDTH = DDSD.WIDTH
|
||||
|
||||
DDS_SURFACE_FLAGS_TEXTURE = DDSCAPS.TEXTURE
|
||||
DDS_SURFACE_FLAGS_MIPMAP = DDSCAPS.COMPLEX | DDSCAPS.MIPMAP
|
||||
DDS_SURFACE_FLAGS_CUBEMAP = DDSCAPS.COMPLEX
|
||||
|
||||
DDS_CUBEMAP_POSITIVEX = DDSCAPS2.CUBEMAP | DDSCAPS2.CUBEMAP_POSITIVEX
|
||||
DDS_CUBEMAP_NEGATIVEX = DDSCAPS2.CUBEMAP | DDSCAPS2.CUBEMAP_NEGATIVEX
|
||||
DDS_CUBEMAP_POSITIVEY = DDSCAPS2.CUBEMAP | DDSCAPS2.CUBEMAP_POSITIVEY
|
||||
DDS_CUBEMAP_NEGATIVEY = DDSCAPS2.CUBEMAP | DDSCAPS2.CUBEMAP_NEGATIVEY
|
||||
DDS_CUBEMAP_POSITIVEZ = DDSCAPS2.CUBEMAP | DDSCAPS2.CUBEMAP_POSITIVEZ
|
||||
DDS_CUBEMAP_NEGATIVEZ = DDSCAPS2.CUBEMAP | DDSCAPS2.CUBEMAP_NEGATIVEZ
|
||||
|
||||
DXT1_FOURCC = D3DFMT.DXT1
|
||||
DXT3_FOURCC = D3DFMT.DXT3
|
||||
DXT5_FOURCC = D3DFMT.DXT5
|
||||
|
||||
DXGI_FORMAT_R8G8B8A8_TYPELESS = DXGI_FORMAT.R8G8B8A8_TYPELESS
|
||||
DXGI_FORMAT_R8G8B8A8_UNORM = DXGI_FORMAT.R8G8B8A8_UNORM
|
||||
DXGI_FORMAT_R8G8B8A8_UNORM_SRGB = DXGI_FORMAT.R8G8B8A8_UNORM_SRGB
|
||||
DXGI_FORMAT_BC5_TYPELESS = DXGI_FORMAT.BC5_TYPELESS
|
||||
DXGI_FORMAT_BC5_UNORM = DXGI_FORMAT.BC5_UNORM
|
||||
DXGI_FORMAT_BC5_SNORM = DXGI_FORMAT.BC5_SNORM
|
||||
DXGI_FORMAT_BC6H_UF16 = DXGI_FORMAT.BC6H_UF16
|
||||
DXGI_FORMAT_BC6H_SF16 = DXGI_FORMAT.BC6H_SF16
|
||||
DXGI_FORMAT_BC7_TYPELESS = DXGI_FORMAT.BC7_TYPELESS
|
||||
DXGI_FORMAT_BC7_UNORM = DXGI_FORMAT.BC7_UNORM
|
||||
DXGI_FORMAT_BC7_UNORM_SRGB = DXGI_FORMAT.BC7_UNORM_SRGB
|
||||
|
||||
|
||||
class DdsImageFile(ImageFile.ImageFile):
|
||||
format = "DDS"
|
||||
format_description = "DirectDraw Surface"
|
||||
|
||||
def _open(self) -> None:
|
||||
if not _accept(self.fp.read(4)):
|
||||
msg = "not a DDS file"
|
||||
raise SyntaxError(msg)
|
||||
(header_size,) = struct.unpack("<I", self.fp.read(4))
|
||||
if header_size != 124:
|
||||
msg = f"Unsupported header size {repr(header_size)}"
|
||||
raise OSError(msg)
|
||||
header_bytes = self.fp.read(header_size - 4)
|
||||
if len(header_bytes) != 120:
|
||||
msg = f"Incomplete header: {len(header_bytes)} bytes"
|
||||
raise OSError(msg)
|
||||
header = io.BytesIO(header_bytes)
|
||||
|
||||
flags, height, width = struct.unpack("<3I", header.read(12))
|
||||
self._size = (width, height)
|
||||
extents = (0, 0) + self.size
|
||||
|
||||
pitch, depth, mipmaps = struct.unpack("<3I", header.read(12))
|
||||
struct.unpack("<11I", header.read(44)) # reserved
|
||||
|
||||
# pixel format
|
||||
pfsize, pfflags, fourcc, bitcount = struct.unpack("<4I", header.read(16))
|
||||
n = 0
|
||||
rawmode = None
|
||||
if pfflags & DDPF.RGB:
|
||||
# Texture contains uncompressed RGB data
|
||||
if pfflags & DDPF.ALPHAPIXELS:
|
||||
self._mode = "RGBA"
|
||||
mask_count = 4
|
||||
else:
|
||||
self._mode = "RGB"
|
||||
mask_count = 3
|
||||
|
||||
masks = struct.unpack(f"<{mask_count}I", header.read(mask_count * 4))
|
||||
self.tile = [ImageFile._Tile("dds_rgb", extents, 0, (bitcount, masks))]
|
||||
return
|
||||
elif pfflags & DDPF.LUMINANCE:
|
||||
if bitcount == 8:
|
||||
self._mode = "L"
|
||||
elif bitcount == 16 and pfflags & DDPF.ALPHAPIXELS:
|
||||
self._mode = "LA"
|
||||
else:
|
||||
msg = f"Unsupported bitcount {bitcount} for {pfflags}"
|
||||
raise OSError(msg)
|
||||
elif pfflags & DDPF.PALETTEINDEXED8:
|
||||
self._mode = "P"
|
||||
self.palette = ImagePalette.raw("RGBA", self.fp.read(1024))
|
||||
self.palette.mode = "RGBA"
|
||||
elif pfflags & DDPF.FOURCC:
|
||||
offset = header_size + 4
|
||||
if fourcc == D3DFMT.DXT1:
|
||||
self._mode = "RGBA"
|
||||
self.pixel_format = "DXT1"
|
||||
n = 1
|
||||
elif fourcc == D3DFMT.DXT3:
|
||||
self._mode = "RGBA"
|
||||
self.pixel_format = "DXT3"
|
||||
n = 2
|
||||
elif fourcc == D3DFMT.DXT5:
|
||||
self._mode = "RGBA"
|
||||
self.pixel_format = "DXT5"
|
||||
n = 3
|
||||
elif fourcc in (D3DFMT.BC4U, D3DFMT.ATI1):
|
||||
self._mode = "L"
|
||||
self.pixel_format = "BC4"
|
||||
n = 4
|
||||
elif fourcc == D3DFMT.BC5S:
|
||||
self._mode = "RGB"
|
||||
self.pixel_format = "BC5S"
|
||||
n = 5
|
||||
elif fourcc in (D3DFMT.BC5U, D3DFMT.ATI2):
|
||||
self._mode = "RGB"
|
||||
self.pixel_format = "BC5"
|
||||
n = 5
|
||||
elif fourcc == D3DFMT.DX10:
|
||||
offset += 20
|
||||
# ignoring flags which pertain to volume textures and cubemaps
|
||||
(dxgi_format,) = struct.unpack("<I", self.fp.read(4))
|
||||
self.fp.read(16)
|
||||
if dxgi_format in (
|
||||
DXGI_FORMAT.BC1_UNORM,
|
||||
DXGI_FORMAT.BC1_TYPELESS,
|
||||
):
|
||||
self._mode = "RGBA"
|
||||
self.pixel_format = "BC1"
|
||||
n = 1
|
||||
elif dxgi_format in (DXGI_FORMAT.BC4_TYPELESS, DXGI_FORMAT.BC4_UNORM):
|
||||
self._mode = "L"
|
||||
self.pixel_format = "BC4"
|
||||
n = 4
|
||||
elif dxgi_format in (DXGI_FORMAT.BC5_TYPELESS, DXGI_FORMAT.BC5_UNORM):
|
||||
self._mode = "RGB"
|
||||
self.pixel_format = "BC5"
|
||||
n = 5
|
||||
elif dxgi_format == DXGI_FORMAT.BC5_SNORM:
|
||||
self._mode = "RGB"
|
||||
self.pixel_format = "BC5S"
|
||||
n = 5
|
||||
elif dxgi_format == DXGI_FORMAT.BC6H_UF16:
|
||||
self._mode = "RGB"
|
||||
self.pixel_format = "BC6H"
|
||||
n = 6
|
||||
elif dxgi_format == DXGI_FORMAT.BC6H_SF16:
|
||||
self._mode = "RGB"
|
||||
self.pixel_format = "BC6HS"
|
||||
n = 6
|
||||
elif dxgi_format in (
|
||||
DXGI_FORMAT.BC7_TYPELESS,
|
||||
DXGI_FORMAT.BC7_UNORM,
|
||||
DXGI_FORMAT.BC7_UNORM_SRGB,
|
||||
):
|
||||
self._mode = "RGBA"
|
||||
self.pixel_format = "BC7"
|
||||
n = 7
|
||||
if dxgi_format == DXGI_FORMAT.BC7_UNORM_SRGB:
|
||||
self.info["gamma"] = 1 / 2.2
|
||||
elif dxgi_format in (
|
||||
DXGI_FORMAT.R8G8B8A8_TYPELESS,
|
||||
DXGI_FORMAT.R8G8B8A8_UNORM,
|
||||
DXGI_FORMAT.R8G8B8A8_UNORM_SRGB,
|
||||
):
|
||||
self._mode = "RGBA"
|
||||
if dxgi_format == DXGI_FORMAT.R8G8B8A8_UNORM_SRGB:
|
||||
self.info["gamma"] = 1 / 2.2
|
||||
else:
|
||||
msg = f"Unimplemented DXGI format {dxgi_format}"
|
||||
raise NotImplementedError(msg)
|
||||
else:
|
||||
msg = f"Unimplemented pixel format {repr(fourcc)}"
|
||||
raise NotImplementedError(msg)
|
||||
else:
|
||||
msg = f"Unknown pixel format flags {pfflags}"
|
||||
raise NotImplementedError(msg)
|
||||
|
||||
if n:
|
||||
self.tile = [
|
||||
ImageFile._Tile("bcn", extents, offset, (n, self.pixel_format))
|
||||
]
|
||||
else:
|
||||
self.tile = [ImageFile._Tile("raw", extents, 0, rawmode or self.mode)]
|
||||
|
||||
def load_seek(self, pos: int) -> None:
|
||||
pass
|
||||
|
||||
|
||||
class DdsRgbDecoder(ImageFile.PyDecoder):
|
||||
_pulls_fd = True
|
||||
|
||||
def decode(self, buffer: bytes | Image.SupportsArrayInterface) -> tuple[int, int]:
|
||||
assert self.fd is not None
|
||||
bitcount, masks = self.args
|
||||
|
||||
# Some masks will be padded with zeros, e.g. R 0b11 G 0b1100
|
||||
# Calculate how many zeros each mask is padded with
|
||||
mask_offsets = []
|
||||
# And the maximum value of each channel without the padding
|
||||
mask_totals = []
|
||||
for mask in masks:
|
||||
offset = 0
|
||||
if mask != 0:
|
||||
while mask >> (offset + 1) << (offset + 1) == mask:
|
||||
offset += 1
|
||||
mask_offsets.append(offset)
|
||||
mask_totals.append(mask >> offset)
|
||||
|
||||
data = bytearray()
|
||||
bytecount = bitcount // 8
|
||||
dest_length = self.state.xsize * self.state.ysize * len(masks)
|
||||
while len(data) < dest_length:
|
||||
value = int.from_bytes(self.fd.read(bytecount), "little")
|
||||
for i, mask in enumerate(masks):
|
||||
masked_value = value & mask
|
||||
# Remove the zero padding, and scale it to 8 bits
|
||||
data += o8(
|
||||
int(((masked_value >> mask_offsets[i]) / mask_totals[i]) * 255)
|
||||
)
|
||||
self.set_as_raw(data)
|
||||
return -1, 0
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
if im.mode not in ("RGB", "RGBA", "L", "LA"):
|
||||
msg = f"cannot write mode {im.mode} as DDS"
|
||||
raise OSError(msg)
|
||||
|
||||
alpha = im.mode[-1] == "A"
|
||||
if im.mode[0] == "L":
|
||||
pixel_flags = DDPF.LUMINANCE
|
||||
rawmode = im.mode
|
||||
if alpha:
|
||||
rgba_mask = [0x000000FF, 0x000000FF, 0x000000FF]
|
||||
else:
|
||||
rgba_mask = [0xFF000000, 0xFF000000, 0xFF000000]
|
||||
else:
|
||||
pixel_flags = DDPF.RGB
|
||||
rawmode = im.mode[::-1]
|
||||
rgba_mask = [0x00FF0000, 0x0000FF00, 0x000000FF]
|
||||
|
||||
if alpha:
|
||||
r, g, b, a = im.split()
|
||||
im = Image.merge("RGBA", (a, r, g, b))
|
||||
if alpha:
|
||||
pixel_flags |= DDPF.ALPHAPIXELS
|
||||
rgba_mask.append(0xFF000000 if alpha else 0)
|
||||
|
||||
flags = DDSD.CAPS | DDSD.HEIGHT | DDSD.WIDTH | DDSD.PITCH | DDSD.PIXELFORMAT
|
||||
bitcount = len(im.getbands()) * 8
|
||||
pitch = (im.width * bitcount + 7) // 8
|
||||
|
||||
fp.write(
|
||||
o32(DDS_MAGIC)
|
||||
+ struct.pack(
|
||||
"<7I",
|
||||
124, # header size
|
||||
flags, # flags
|
||||
im.height,
|
||||
im.width,
|
||||
pitch,
|
||||
0, # depth
|
||||
0, # mipmaps
|
||||
)
|
||||
+ struct.pack("11I", *((0,) * 11)) # reserved
|
||||
# pfsize, pfflags, fourcc, bitcount
|
||||
+ struct.pack("<4I", 32, pixel_flags, 0, bitcount)
|
||||
+ struct.pack("<4I", *rgba_mask) # dwRGBABitMask
|
||||
+ struct.pack("<5I", DDSCAPS.TEXTURE, 0, 0, 0, 0)
|
||||
)
|
||||
ImageFile._save(im, fp, [ImageFile._Tile("raw", (0, 0) + im.size, 0, rawmode)])
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == b"DDS "
|
||||
|
||||
|
||||
Image.register_open(DdsImageFile.format, DdsImageFile, _accept)
|
||||
Image.register_decoder("dds_rgb", DdsRgbDecoder)
|
||||
Image.register_save(DdsImageFile.format, _save)
|
||||
Image.register_extension(DdsImageFile.format, ".dds")
|
||||
474
testclass/lib/python3.12/site-packages/PIL/EpsImagePlugin.py
Normal file
474
testclass/lib/python3.12/site-packages/PIL/EpsImagePlugin.py
Normal file
@ -0,0 +1,474 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# EPS file handling
|
||||
#
|
||||
# History:
|
||||
# 1995-09-01 fl Created (0.1)
|
||||
# 1996-05-18 fl Don't choke on "atend" fields, Ghostscript interface (0.2)
|
||||
# 1996-08-22 fl Don't choke on floating point BoundingBox values
|
||||
# 1996-08-23 fl Handle files from Macintosh (0.3)
|
||||
# 2001-02-17 fl Use 're' instead of 'regex' (Python 2.1) (0.4)
|
||||
# 2003-09-07 fl Check gs.close status (from Federico Di Gregorio) (0.5)
|
||||
# 2014-05-07 e Handling of EPS with binary preview and fixed resolution
|
||||
# resizing
|
||||
#
|
||||
# Copyright (c) 1997-2003 by Secret Labs AB.
|
||||
# Copyright (c) 1995-2003 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import io
|
||||
import os
|
||||
import re
|
||||
import subprocess
|
||||
import sys
|
||||
import tempfile
|
||||
from typing import IO
|
||||
|
||||
from . import Image, ImageFile
|
||||
from ._binary import i32le as i32
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
|
||||
split = re.compile(r"^%%([^:]*):[ \t]*(.*)[ \t]*$")
|
||||
field = re.compile(r"^%[%!\w]([^:]*)[ \t]*$")
|
||||
|
||||
gs_binary: str | bool | None = None
|
||||
gs_windows_binary = None
|
||||
|
||||
|
||||
def has_ghostscript() -> bool:
|
||||
global gs_binary, gs_windows_binary
|
||||
if gs_binary is None:
|
||||
if sys.platform.startswith("win"):
|
||||
if gs_windows_binary is None:
|
||||
import shutil
|
||||
|
||||
for binary in ("gswin32c", "gswin64c", "gs"):
|
||||
if shutil.which(binary) is not None:
|
||||
gs_windows_binary = binary
|
||||
break
|
||||
else:
|
||||
gs_windows_binary = False
|
||||
gs_binary = gs_windows_binary
|
||||
else:
|
||||
try:
|
||||
subprocess.check_call(["gs", "--version"], stdout=subprocess.DEVNULL)
|
||||
gs_binary = "gs"
|
||||
except OSError:
|
||||
gs_binary = False
|
||||
return gs_binary is not False
|
||||
|
||||
|
||||
def Ghostscript(
|
||||
tile: list[ImageFile._Tile],
|
||||
size: tuple[int, int],
|
||||
fp: IO[bytes],
|
||||
scale: int = 1,
|
||||
transparency: bool = False,
|
||||
) -> Image.core.ImagingCore:
|
||||
"""Render an image using Ghostscript"""
|
||||
global gs_binary
|
||||
if not has_ghostscript():
|
||||
msg = "Unable to locate Ghostscript on paths"
|
||||
raise OSError(msg)
|
||||
assert isinstance(gs_binary, str)
|
||||
|
||||
# Unpack decoder tile
|
||||
args = tile[0].args
|
||||
assert isinstance(args, tuple)
|
||||
length, bbox = args
|
||||
|
||||
# Hack to support hi-res rendering
|
||||
scale = int(scale) or 1
|
||||
width = size[0] * scale
|
||||
height = size[1] * scale
|
||||
# resolution is dependent on bbox and size
|
||||
res_x = 72.0 * width / (bbox[2] - bbox[0])
|
||||
res_y = 72.0 * height / (bbox[3] - bbox[1])
|
||||
|
||||
out_fd, outfile = tempfile.mkstemp()
|
||||
os.close(out_fd)
|
||||
|
||||
infile_temp = None
|
||||
if hasattr(fp, "name") and os.path.exists(fp.name):
|
||||
infile = fp.name
|
||||
else:
|
||||
in_fd, infile_temp = tempfile.mkstemp()
|
||||
os.close(in_fd)
|
||||
infile = infile_temp
|
||||
|
||||
# Ignore length and offset!
|
||||
# Ghostscript can read it
|
||||
# Copy whole file to read in Ghostscript
|
||||
with open(infile_temp, "wb") as f:
|
||||
# fetch length of fp
|
||||
fp.seek(0, io.SEEK_END)
|
||||
fsize = fp.tell()
|
||||
# ensure start position
|
||||
# go back
|
||||
fp.seek(0)
|
||||
lengthfile = fsize
|
||||
while lengthfile > 0:
|
||||
s = fp.read(min(lengthfile, 100 * 1024))
|
||||
if not s:
|
||||
break
|
||||
lengthfile -= len(s)
|
||||
f.write(s)
|
||||
|
||||
if transparency:
|
||||
# "RGBA"
|
||||
device = "pngalpha"
|
||||
else:
|
||||
# "pnmraw" automatically chooses between
|
||||
# PBM ("1"), PGM ("L"), and PPM ("RGB").
|
||||
device = "pnmraw"
|
||||
|
||||
# Build Ghostscript command
|
||||
command = [
|
||||
gs_binary,
|
||||
"-q", # quiet mode
|
||||
f"-g{width:d}x{height:d}", # set output geometry (pixels)
|
||||
f"-r{res_x:f}x{res_y:f}", # set input DPI (dots per inch)
|
||||
"-dBATCH", # exit after processing
|
||||
"-dNOPAUSE", # don't pause between pages
|
||||
"-dSAFER", # safe mode
|
||||
f"-sDEVICE={device}",
|
||||
f"-sOutputFile={outfile}", # output file
|
||||
# adjust for image origin
|
||||
"-c",
|
||||
f"{-bbox[0]} {-bbox[1]} translate",
|
||||
"-f",
|
||||
infile, # input file
|
||||
# showpage (see https://bugs.ghostscript.com/show_bug.cgi?id=698272)
|
||||
"-c",
|
||||
"showpage",
|
||||
]
|
||||
|
||||
# push data through Ghostscript
|
||||
try:
|
||||
startupinfo = None
|
||||
if sys.platform.startswith("win"):
|
||||
startupinfo = subprocess.STARTUPINFO()
|
||||
startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW
|
||||
subprocess.check_call(command, startupinfo=startupinfo)
|
||||
with Image.open(outfile) as out_im:
|
||||
out_im.load()
|
||||
return out_im.im.copy()
|
||||
finally:
|
||||
try:
|
||||
os.unlink(outfile)
|
||||
if infile_temp:
|
||||
os.unlink(infile_temp)
|
||||
except OSError:
|
||||
pass
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == b"%!PS" or (len(prefix) >= 4 and i32(prefix) == 0xC6D3D0C5)
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for Encapsulated PostScript. This plugin supports only
|
||||
# a few variants of this format.
|
||||
|
||||
|
||||
class EpsImageFile(ImageFile.ImageFile):
|
||||
"""EPS File Parser for the Python Imaging Library"""
|
||||
|
||||
format = "EPS"
|
||||
format_description = "Encapsulated Postscript"
|
||||
|
||||
mode_map = {1: "L", 2: "LAB", 3: "RGB", 4: "CMYK"}
|
||||
|
||||
def _open(self) -> None:
|
||||
(length, offset) = self._find_offset(self.fp)
|
||||
|
||||
# go to offset - start of "%!PS"
|
||||
self.fp.seek(offset)
|
||||
|
||||
self._mode = "RGB"
|
||||
|
||||
# When reading header comments, the first comment is used.
|
||||
# When reading trailer comments, the last comment is used.
|
||||
bounding_box: list[int] | None = None
|
||||
imagedata_size: tuple[int, int] | None = None
|
||||
|
||||
byte_arr = bytearray(255)
|
||||
bytes_mv = memoryview(byte_arr)
|
||||
bytes_read = 0
|
||||
reading_header_comments = True
|
||||
reading_trailer_comments = False
|
||||
trailer_reached = False
|
||||
|
||||
def check_required_header_comments() -> None:
|
||||
"""
|
||||
The EPS specification requires that some headers exist.
|
||||
This should be checked when the header comments formally end,
|
||||
when image data starts, or when the file ends, whichever comes first.
|
||||
"""
|
||||
if "PS-Adobe" not in self.info:
|
||||
msg = 'EPS header missing "%!PS-Adobe" comment'
|
||||
raise SyntaxError(msg)
|
||||
if "BoundingBox" not in self.info:
|
||||
msg = 'EPS header missing "%%BoundingBox" comment'
|
||||
raise SyntaxError(msg)
|
||||
|
||||
def read_comment(s: str) -> bool:
|
||||
nonlocal bounding_box, reading_trailer_comments
|
||||
try:
|
||||
m = split.match(s)
|
||||
except re.error as e:
|
||||
msg = "not an EPS file"
|
||||
raise SyntaxError(msg) from e
|
||||
|
||||
if not m:
|
||||
return False
|
||||
|
||||
k, v = m.group(1, 2)
|
||||
self.info[k] = v
|
||||
if k == "BoundingBox":
|
||||
if v == "(atend)":
|
||||
reading_trailer_comments = True
|
||||
elif not bounding_box or (trailer_reached and reading_trailer_comments):
|
||||
try:
|
||||
# Note: The DSC spec says that BoundingBox
|
||||
# fields should be integers, but some drivers
|
||||
# put floating point values there anyway.
|
||||
bounding_box = [int(float(i)) for i in v.split()]
|
||||
except Exception:
|
||||
pass
|
||||
return True
|
||||
|
||||
while True:
|
||||
byte = self.fp.read(1)
|
||||
if byte == b"":
|
||||
# if we didn't read a byte we must be at the end of the file
|
||||
if bytes_read == 0:
|
||||
if reading_header_comments:
|
||||
check_required_header_comments()
|
||||
break
|
||||
elif byte in b"\r\n":
|
||||
# if we read a line ending character, ignore it and parse what
|
||||
# we have already read. if we haven't read any other characters,
|
||||
# continue reading
|
||||
if bytes_read == 0:
|
||||
continue
|
||||
else:
|
||||
# ASCII/hexadecimal lines in an EPS file must not exceed
|
||||
# 255 characters, not including line ending characters
|
||||
if bytes_read >= 255:
|
||||
# only enforce this for lines starting with a "%",
|
||||
# otherwise assume it's binary data
|
||||
if byte_arr[0] == ord("%"):
|
||||
msg = "not an EPS file"
|
||||
raise SyntaxError(msg)
|
||||
else:
|
||||
if reading_header_comments:
|
||||
check_required_header_comments()
|
||||
reading_header_comments = False
|
||||
# reset bytes_read so we can keep reading
|
||||
# data until the end of the line
|
||||
bytes_read = 0
|
||||
byte_arr[bytes_read] = byte[0]
|
||||
bytes_read += 1
|
||||
continue
|
||||
|
||||
if reading_header_comments:
|
||||
# Load EPS header
|
||||
|
||||
# if this line doesn't start with a "%",
|
||||
# or does start with "%%EndComments",
|
||||
# then we've reached the end of the header/comments
|
||||
if byte_arr[0] != ord("%") or bytes_mv[:13] == b"%%EndComments":
|
||||
check_required_header_comments()
|
||||
reading_header_comments = False
|
||||
continue
|
||||
|
||||
s = str(bytes_mv[:bytes_read], "latin-1")
|
||||
if not read_comment(s):
|
||||
m = field.match(s)
|
||||
if m:
|
||||
k = m.group(1)
|
||||
if k[:8] == "PS-Adobe":
|
||||
self.info["PS-Adobe"] = k[9:]
|
||||
else:
|
||||
self.info[k] = ""
|
||||
elif s[0] == "%":
|
||||
# handle non-DSC PostScript comments that some
|
||||
# tools mistakenly put in the Comments section
|
||||
pass
|
||||
else:
|
||||
msg = "bad EPS header"
|
||||
raise OSError(msg)
|
||||
elif bytes_mv[:11] == b"%ImageData:":
|
||||
# Check for an "ImageData" descriptor
|
||||
# https://www.adobe.com/devnet-apps/photoshop/fileformatashtml/#50577413_pgfId-1035096
|
||||
|
||||
# If we've already read an "ImageData" descriptor,
|
||||
# don't read another one.
|
||||
if imagedata_size:
|
||||
bytes_read = 0
|
||||
continue
|
||||
|
||||
# Values:
|
||||
# columns
|
||||
# rows
|
||||
# bit depth (1 or 8)
|
||||
# mode (1: L, 2: LAB, 3: RGB, 4: CMYK)
|
||||
# number of padding channels
|
||||
# block size (number of bytes per row per channel)
|
||||
# binary/ascii (1: binary, 2: ascii)
|
||||
# data start identifier (the image data follows after a single line
|
||||
# consisting only of this quoted value)
|
||||
image_data_values = byte_arr[11:bytes_read].split(None, 7)
|
||||
columns, rows, bit_depth, mode_id = (
|
||||
int(value) for value in image_data_values[:4]
|
||||
)
|
||||
|
||||
if bit_depth == 1:
|
||||
self._mode = "1"
|
||||
elif bit_depth == 8:
|
||||
try:
|
||||
self._mode = self.mode_map[mode_id]
|
||||
except ValueError:
|
||||
break
|
||||
else:
|
||||
break
|
||||
|
||||
# Parse the columns and rows after checking the bit depth and mode
|
||||
# in case the bit depth and/or mode are invalid.
|
||||
imagedata_size = columns, rows
|
||||
elif bytes_mv[:5] == b"%%EOF":
|
||||
break
|
||||
elif trailer_reached and reading_trailer_comments:
|
||||
# Load EPS trailer
|
||||
s = str(bytes_mv[:bytes_read], "latin-1")
|
||||
read_comment(s)
|
||||
elif bytes_mv[:9] == b"%%Trailer":
|
||||
trailer_reached = True
|
||||
bytes_read = 0
|
||||
|
||||
# A "BoundingBox" is always required,
|
||||
# even if an "ImageData" descriptor size exists.
|
||||
if not bounding_box:
|
||||
msg = "cannot determine EPS bounding box"
|
||||
raise OSError(msg)
|
||||
|
||||
# An "ImageData" size takes precedence over the "BoundingBox".
|
||||
self._size = imagedata_size or (
|
||||
bounding_box[2] - bounding_box[0],
|
||||
bounding_box[3] - bounding_box[1],
|
||||
)
|
||||
|
||||
self.tile = [
|
||||
ImageFile._Tile("eps", (0, 0) + self.size, offset, (length, bounding_box))
|
||||
]
|
||||
|
||||
def _find_offset(self, fp: IO[bytes]) -> tuple[int, int]:
|
||||
s = fp.read(4)
|
||||
|
||||
if s == b"%!PS":
|
||||
# for HEAD without binary preview
|
||||
fp.seek(0, io.SEEK_END)
|
||||
length = fp.tell()
|
||||
offset = 0
|
||||
elif i32(s) == 0xC6D3D0C5:
|
||||
# FIX for: Some EPS file not handled correctly / issue #302
|
||||
# EPS can contain binary data
|
||||
# or start directly with latin coding
|
||||
# more info see:
|
||||
# https://web.archive.org/web/20160528181353/http://partners.adobe.com/public/developer/en/ps/5002.EPSF_Spec.pdf
|
||||
s = fp.read(8)
|
||||
offset = i32(s)
|
||||
length = i32(s, 4)
|
||||
else:
|
||||
msg = "not an EPS file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
return length, offset
|
||||
|
||||
def load(
|
||||
self, scale: int = 1, transparency: bool = False
|
||||
) -> Image.core.PixelAccess | None:
|
||||
# Load EPS via Ghostscript
|
||||
if self.tile:
|
||||
self.im = Ghostscript(self.tile, self.size, self.fp, scale, transparency)
|
||||
self._mode = self.im.mode
|
||||
self._size = self.im.size
|
||||
self.tile = []
|
||||
return Image.Image.load(self)
|
||||
|
||||
def load_seek(self, pos: int) -> None:
|
||||
# we can't incrementally load, so force ImageFile.parser to
|
||||
# use our custom load method by defining this method.
|
||||
pass
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes, eps: int = 1) -> None:
|
||||
"""EPS Writer for the Python Imaging Library."""
|
||||
|
||||
# make sure image data is available
|
||||
im.load()
|
||||
|
||||
# determine PostScript image mode
|
||||
if im.mode == "L":
|
||||
operator = (8, 1, b"image")
|
||||
elif im.mode == "RGB":
|
||||
operator = (8, 3, b"false 3 colorimage")
|
||||
elif im.mode == "CMYK":
|
||||
operator = (8, 4, b"false 4 colorimage")
|
||||
else:
|
||||
msg = "image mode is not supported"
|
||||
raise ValueError(msg)
|
||||
|
||||
if eps:
|
||||
# write EPS header
|
||||
fp.write(b"%!PS-Adobe-3.0 EPSF-3.0\n")
|
||||
fp.write(b"%%Creator: PIL 0.1 EpsEncode\n")
|
||||
# fp.write("%%CreationDate: %s"...)
|
||||
fp.write(b"%%%%BoundingBox: 0 0 %d %d\n" % im.size)
|
||||
fp.write(b"%%Pages: 1\n")
|
||||
fp.write(b"%%EndComments\n")
|
||||
fp.write(b"%%Page: 1 1\n")
|
||||
fp.write(b"%%ImageData: %d %d " % im.size)
|
||||
fp.write(b'%d %d 0 1 1 "%s"\n' % operator)
|
||||
|
||||
# image header
|
||||
fp.write(b"gsave\n")
|
||||
fp.write(b"10 dict begin\n")
|
||||
fp.write(b"/buf %d string def\n" % (im.size[0] * operator[1]))
|
||||
fp.write(b"%d %d scale\n" % im.size)
|
||||
fp.write(b"%d %d 8\n" % im.size) # <= bits
|
||||
fp.write(b"[%d 0 0 -%d 0 %d]\n" % (im.size[0], im.size[1], im.size[1]))
|
||||
fp.write(b"{ currentfile buf readhexstring pop } bind\n")
|
||||
fp.write(operator[2] + b"\n")
|
||||
if hasattr(fp, "flush"):
|
||||
fp.flush()
|
||||
|
||||
ImageFile._save(im, fp, [ImageFile._Tile("eps", (0, 0) + im.size)])
|
||||
|
||||
fp.write(b"\n%%%%EndBinary\n")
|
||||
fp.write(b"grestore end\n")
|
||||
if hasattr(fp, "flush"):
|
||||
fp.flush()
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
|
||||
Image.register_open(EpsImageFile.format, EpsImageFile, _accept)
|
||||
|
||||
Image.register_save(EpsImageFile.format, _save)
|
||||
|
||||
Image.register_extensions(EpsImageFile.format, [".ps", ".eps"])
|
||||
|
||||
Image.register_mime(EpsImageFile.format, "application/postscript")
|
||||
382
testclass/lib/python3.12/site-packages/PIL/ExifTags.py
Normal file
382
testclass/lib/python3.12/site-packages/PIL/ExifTags.py
Normal file
@ -0,0 +1,382 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# EXIF tags
|
||||
#
|
||||
# Copyright (c) 2003 by Secret Labs AB
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
|
||||
"""
|
||||
This module provides constants and clear-text names for various
|
||||
well-known EXIF tags.
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
from enum import IntEnum
|
||||
|
||||
|
||||
class Base(IntEnum):
|
||||
# possibly incomplete
|
||||
InteropIndex = 0x0001
|
||||
ProcessingSoftware = 0x000B
|
||||
NewSubfileType = 0x00FE
|
||||
SubfileType = 0x00FF
|
||||
ImageWidth = 0x0100
|
||||
ImageLength = 0x0101
|
||||
BitsPerSample = 0x0102
|
||||
Compression = 0x0103
|
||||
PhotometricInterpretation = 0x0106
|
||||
Thresholding = 0x0107
|
||||
CellWidth = 0x0108
|
||||
CellLength = 0x0109
|
||||
FillOrder = 0x010A
|
||||
DocumentName = 0x010D
|
||||
ImageDescription = 0x010E
|
||||
Make = 0x010F
|
||||
Model = 0x0110
|
||||
StripOffsets = 0x0111
|
||||
Orientation = 0x0112
|
||||
SamplesPerPixel = 0x0115
|
||||
RowsPerStrip = 0x0116
|
||||
StripByteCounts = 0x0117
|
||||
MinSampleValue = 0x0118
|
||||
MaxSampleValue = 0x0119
|
||||
XResolution = 0x011A
|
||||
YResolution = 0x011B
|
||||
PlanarConfiguration = 0x011C
|
||||
PageName = 0x011D
|
||||
FreeOffsets = 0x0120
|
||||
FreeByteCounts = 0x0121
|
||||
GrayResponseUnit = 0x0122
|
||||
GrayResponseCurve = 0x0123
|
||||
T4Options = 0x0124
|
||||
T6Options = 0x0125
|
||||
ResolutionUnit = 0x0128
|
||||
PageNumber = 0x0129
|
||||
TransferFunction = 0x012D
|
||||
Software = 0x0131
|
||||
DateTime = 0x0132
|
||||
Artist = 0x013B
|
||||
HostComputer = 0x013C
|
||||
Predictor = 0x013D
|
||||
WhitePoint = 0x013E
|
||||
PrimaryChromaticities = 0x013F
|
||||
ColorMap = 0x0140
|
||||
HalftoneHints = 0x0141
|
||||
TileWidth = 0x0142
|
||||
TileLength = 0x0143
|
||||
TileOffsets = 0x0144
|
||||
TileByteCounts = 0x0145
|
||||
SubIFDs = 0x014A
|
||||
InkSet = 0x014C
|
||||
InkNames = 0x014D
|
||||
NumberOfInks = 0x014E
|
||||
DotRange = 0x0150
|
||||
TargetPrinter = 0x0151
|
||||
ExtraSamples = 0x0152
|
||||
SampleFormat = 0x0153
|
||||
SMinSampleValue = 0x0154
|
||||
SMaxSampleValue = 0x0155
|
||||
TransferRange = 0x0156
|
||||
ClipPath = 0x0157
|
||||
XClipPathUnits = 0x0158
|
||||
YClipPathUnits = 0x0159
|
||||
Indexed = 0x015A
|
||||
JPEGTables = 0x015B
|
||||
OPIProxy = 0x015F
|
||||
JPEGProc = 0x0200
|
||||
JpegIFOffset = 0x0201
|
||||
JpegIFByteCount = 0x0202
|
||||
JpegRestartInterval = 0x0203
|
||||
JpegLosslessPredictors = 0x0205
|
||||
JpegPointTransforms = 0x0206
|
||||
JpegQTables = 0x0207
|
||||
JpegDCTables = 0x0208
|
||||
JpegACTables = 0x0209
|
||||
YCbCrCoefficients = 0x0211
|
||||
YCbCrSubSampling = 0x0212
|
||||
YCbCrPositioning = 0x0213
|
||||
ReferenceBlackWhite = 0x0214
|
||||
XMLPacket = 0x02BC
|
||||
RelatedImageFileFormat = 0x1000
|
||||
RelatedImageWidth = 0x1001
|
||||
RelatedImageLength = 0x1002
|
||||
Rating = 0x4746
|
||||
RatingPercent = 0x4749
|
||||
ImageID = 0x800D
|
||||
CFARepeatPatternDim = 0x828D
|
||||
BatteryLevel = 0x828F
|
||||
Copyright = 0x8298
|
||||
ExposureTime = 0x829A
|
||||
FNumber = 0x829D
|
||||
IPTCNAA = 0x83BB
|
||||
ImageResources = 0x8649
|
||||
ExifOffset = 0x8769
|
||||
InterColorProfile = 0x8773
|
||||
ExposureProgram = 0x8822
|
||||
SpectralSensitivity = 0x8824
|
||||
GPSInfo = 0x8825
|
||||
ISOSpeedRatings = 0x8827
|
||||
OECF = 0x8828
|
||||
Interlace = 0x8829
|
||||
TimeZoneOffset = 0x882A
|
||||
SelfTimerMode = 0x882B
|
||||
SensitivityType = 0x8830
|
||||
StandardOutputSensitivity = 0x8831
|
||||
RecommendedExposureIndex = 0x8832
|
||||
ISOSpeed = 0x8833
|
||||
ISOSpeedLatitudeyyy = 0x8834
|
||||
ISOSpeedLatitudezzz = 0x8835
|
||||
ExifVersion = 0x9000
|
||||
DateTimeOriginal = 0x9003
|
||||
DateTimeDigitized = 0x9004
|
||||
OffsetTime = 0x9010
|
||||
OffsetTimeOriginal = 0x9011
|
||||
OffsetTimeDigitized = 0x9012
|
||||
ComponentsConfiguration = 0x9101
|
||||
CompressedBitsPerPixel = 0x9102
|
||||
ShutterSpeedValue = 0x9201
|
||||
ApertureValue = 0x9202
|
||||
BrightnessValue = 0x9203
|
||||
ExposureBiasValue = 0x9204
|
||||
MaxApertureValue = 0x9205
|
||||
SubjectDistance = 0x9206
|
||||
MeteringMode = 0x9207
|
||||
LightSource = 0x9208
|
||||
Flash = 0x9209
|
||||
FocalLength = 0x920A
|
||||
Noise = 0x920D
|
||||
ImageNumber = 0x9211
|
||||
SecurityClassification = 0x9212
|
||||
ImageHistory = 0x9213
|
||||
TIFFEPStandardID = 0x9216
|
||||
MakerNote = 0x927C
|
||||
UserComment = 0x9286
|
||||
SubsecTime = 0x9290
|
||||
SubsecTimeOriginal = 0x9291
|
||||
SubsecTimeDigitized = 0x9292
|
||||
AmbientTemperature = 0x9400
|
||||
Humidity = 0x9401
|
||||
Pressure = 0x9402
|
||||
WaterDepth = 0x9403
|
||||
Acceleration = 0x9404
|
||||
CameraElevationAngle = 0x9405
|
||||
XPTitle = 0x9C9B
|
||||
XPComment = 0x9C9C
|
||||
XPAuthor = 0x9C9D
|
||||
XPKeywords = 0x9C9E
|
||||
XPSubject = 0x9C9F
|
||||
FlashPixVersion = 0xA000
|
||||
ColorSpace = 0xA001
|
||||
ExifImageWidth = 0xA002
|
||||
ExifImageHeight = 0xA003
|
||||
RelatedSoundFile = 0xA004
|
||||
ExifInteroperabilityOffset = 0xA005
|
||||
FlashEnergy = 0xA20B
|
||||
SpatialFrequencyResponse = 0xA20C
|
||||
FocalPlaneXResolution = 0xA20E
|
||||
FocalPlaneYResolution = 0xA20F
|
||||
FocalPlaneResolutionUnit = 0xA210
|
||||
SubjectLocation = 0xA214
|
||||
ExposureIndex = 0xA215
|
||||
SensingMethod = 0xA217
|
||||
FileSource = 0xA300
|
||||
SceneType = 0xA301
|
||||
CFAPattern = 0xA302
|
||||
CustomRendered = 0xA401
|
||||
ExposureMode = 0xA402
|
||||
WhiteBalance = 0xA403
|
||||
DigitalZoomRatio = 0xA404
|
||||
FocalLengthIn35mmFilm = 0xA405
|
||||
SceneCaptureType = 0xA406
|
||||
GainControl = 0xA407
|
||||
Contrast = 0xA408
|
||||
Saturation = 0xA409
|
||||
Sharpness = 0xA40A
|
||||
DeviceSettingDescription = 0xA40B
|
||||
SubjectDistanceRange = 0xA40C
|
||||
ImageUniqueID = 0xA420
|
||||
CameraOwnerName = 0xA430
|
||||
BodySerialNumber = 0xA431
|
||||
LensSpecification = 0xA432
|
||||
LensMake = 0xA433
|
||||
LensModel = 0xA434
|
||||
LensSerialNumber = 0xA435
|
||||
CompositeImage = 0xA460
|
||||
CompositeImageCount = 0xA461
|
||||
CompositeImageExposureTimes = 0xA462
|
||||
Gamma = 0xA500
|
||||
PrintImageMatching = 0xC4A5
|
||||
DNGVersion = 0xC612
|
||||
DNGBackwardVersion = 0xC613
|
||||
UniqueCameraModel = 0xC614
|
||||
LocalizedCameraModel = 0xC615
|
||||
CFAPlaneColor = 0xC616
|
||||
CFALayout = 0xC617
|
||||
LinearizationTable = 0xC618
|
||||
BlackLevelRepeatDim = 0xC619
|
||||
BlackLevel = 0xC61A
|
||||
BlackLevelDeltaH = 0xC61B
|
||||
BlackLevelDeltaV = 0xC61C
|
||||
WhiteLevel = 0xC61D
|
||||
DefaultScale = 0xC61E
|
||||
DefaultCropOrigin = 0xC61F
|
||||
DefaultCropSize = 0xC620
|
||||
ColorMatrix1 = 0xC621
|
||||
ColorMatrix2 = 0xC622
|
||||
CameraCalibration1 = 0xC623
|
||||
CameraCalibration2 = 0xC624
|
||||
ReductionMatrix1 = 0xC625
|
||||
ReductionMatrix2 = 0xC626
|
||||
AnalogBalance = 0xC627
|
||||
AsShotNeutral = 0xC628
|
||||
AsShotWhiteXY = 0xC629
|
||||
BaselineExposure = 0xC62A
|
||||
BaselineNoise = 0xC62B
|
||||
BaselineSharpness = 0xC62C
|
||||
BayerGreenSplit = 0xC62D
|
||||
LinearResponseLimit = 0xC62E
|
||||
CameraSerialNumber = 0xC62F
|
||||
LensInfo = 0xC630
|
||||
ChromaBlurRadius = 0xC631
|
||||
AntiAliasStrength = 0xC632
|
||||
ShadowScale = 0xC633
|
||||
DNGPrivateData = 0xC634
|
||||
MakerNoteSafety = 0xC635
|
||||
CalibrationIlluminant1 = 0xC65A
|
||||
CalibrationIlluminant2 = 0xC65B
|
||||
BestQualityScale = 0xC65C
|
||||
RawDataUniqueID = 0xC65D
|
||||
OriginalRawFileName = 0xC68B
|
||||
OriginalRawFileData = 0xC68C
|
||||
ActiveArea = 0xC68D
|
||||
MaskedAreas = 0xC68E
|
||||
AsShotICCProfile = 0xC68F
|
||||
AsShotPreProfileMatrix = 0xC690
|
||||
CurrentICCProfile = 0xC691
|
||||
CurrentPreProfileMatrix = 0xC692
|
||||
ColorimetricReference = 0xC6BF
|
||||
CameraCalibrationSignature = 0xC6F3
|
||||
ProfileCalibrationSignature = 0xC6F4
|
||||
AsShotProfileName = 0xC6F6
|
||||
NoiseReductionApplied = 0xC6F7
|
||||
ProfileName = 0xC6F8
|
||||
ProfileHueSatMapDims = 0xC6F9
|
||||
ProfileHueSatMapData1 = 0xC6FA
|
||||
ProfileHueSatMapData2 = 0xC6FB
|
||||
ProfileToneCurve = 0xC6FC
|
||||
ProfileEmbedPolicy = 0xC6FD
|
||||
ProfileCopyright = 0xC6FE
|
||||
ForwardMatrix1 = 0xC714
|
||||
ForwardMatrix2 = 0xC715
|
||||
PreviewApplicationName = 0xC716
|
||||
PreviewApplicationVersion = 0xC717
|
||||
PreviewSettingsName = 0xC718
|
||||
PreviewSettingsDigest = 0xC719
|
||||
PreviewColorSpace = 0xC71A
|
||||
PreviewDateTime = 0xC71B
|
||||
RawImageDigest = 0xC71C
|
||||
OriginalRawFileDigest = 0xC71D
|
||||
SubTileBlockSize = 0xC71E
|
||||
RowInterleaveFactor = 0xC71F
|
||||
ProfileLookTableDims = 0xC725
|
||||
ProfileLookTableData = 0xC726
|
||||
OpcodeList1 = 0xC740
|
||||
OpcodeList2 = 0xC741
|
||||
OpcodeList3 = 0xC74E
|
||||
NoiseProfile = 0xC761
|
||||
|
||||
|
||||
"""Maps EXIF tags to tag names."""
|
||||
TAGS = {
|
||||
**{i.value: i.name for i in Base},
|
||||
0x920C: "SpatialFrequencyResponse",
|
||||
0x9214: "SubjectLocation",
|
||||
0x9215: "ExposureIndex",
|
||||
0x828E: "CFAPattern",
|
||||
0x920B: "FlashEnergy",
|
||||
0x9216: "TIFF/EPStandardID",
|
||||
}
|
||||
|
||||
|
||||
class GPS(IntEnum):
|
||||
GPSVersionID = 0x00
|
||||
GPSLatitudeRef = 0x01
|
||||
GPSLatitude = 0x02
|
||||
GPSLongitudeRef = 0x03
|
||||
GPSLongitude = 0x04
|
||||
GPSAltitudeRef = 0x05
|
||||
GPSAltitude = 0x06
|
||||
GPSTimeStamp = 0x07
|
||||
GPSSatellites = 0x08
|
||||
GPSStatus = 0x09
|
||||
GPSMeasureMode = 0x0A
|
||||
GPSDOP = 0x0B
|
||||
GPSSpeedRef = 0x0C
|
||||
GPSSpeed = 0x0D
|
||||
GPSTrackRef = 0x0E
|
||||
GPSTrack = 0x0F
|
||||
GPSImgDirectionRef = 0x10
|
||||
GPSImgDirection = 0x11
|
||||
GPSMapDatum = 0x12
|
||||
GPSDestLatitudeRef = 0x13
|
||||
GPSDestLatitude = 0x14
|
||||
GPSDestLongitudeRef = 0x15
|
||||
GPSDestLongitude = 0x16
|
||||
GPSDestBearingRef = 0x17
|
||||
GPSDestBearing = 0x18
|
||||
GPSDestDistanceRef = 0x19
|
||||
GPSDestDistance = 0x1A
|
||||
GPSProcessingMethod = 0x1B
|
||||
GPSAreaInformation = 0x1C
|
||||
GPSDateStamp = 0x1D
|
||||
GPSDifferential = 0x1E
|
||||
GPSHPositioningError = 0x1F
|
||||
|
||||
|
||||
"""Maps EXIF GPS tags to tag names."""
|
||||
GPSTAGS = {i.value: i.name for i in GPS}
|
||||
|
||||
|
||||
class Interop(IntEnum):
|
||||
InteropIndex = 0x0001
|
||||
InteropVersion = 0x0002
|
||||
RelatedImageFileFormat = 0x1000
|
||||
RelatedImageWidth = 0x1001
|
||||
RelatedImageHeight = 0x1002
|
||||
|
||||
|
||||
class IFD(IntEnum):
|
||||
Exif = 0x8769
|
||||
GPSInfo = 0x8825
|
||||
MakerNote = 0x927C
|
||||
Makernote = 0x927C # Deprecated
|
||||
Interop = 0xA005
|
||||
IFD1 = -1
|
||||
|
||||
|
||||
class LightSource(IntEnum):
|
||||
Unknown = 0x00
|
||||
Daylight = 0x01
|
||||
Fluorescent = 0x02
|
||||
Tungsten = 0x03
|
||||
Flash = 0x04
|
||||
Fine = 0x09
|
||||
Cloudy = 0x0A
|
||||
Shade = 0x0B
|
||||
DaylightFluorescent = 0x0C
|
||||
DayWhiteFluorescent = 0x0D
|
||||
CoolWhiteFluorescent = 0x0E
|
||||
WhiteFluorescent = 0x0F
|
||||
StandardLightA = 0x11
|
||||
StandardLightB = 0x12
|
||||
StandardLightC = 0x13
|
||||
D55 = 0x14
|
||||
D65 = 0x15
|
||||
D75 = 0x16
|
||||
D50 = 0x17
|
||||
ISO = 0x18
|
||||
Other = 0xFF
|
||||
152
testclass/lib/python3.12/site-packages/PIL/FitsImagePlugin.py
Normal file
152
testclass/lib/python3.12/site-packages/PIL/FitsImagePlugin.py
Normal file
@ -0,0 +1,152 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# FITS file handling
|
||||
#
|
||||
# Copyright (c) 1998-2003 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import gzip
|
||||
import math
|
||||
|
||||
from . import Image, ImageFile
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:6] == b"SIMPLE"
|
||||
|
||||
|
||||
class FitsImageFile(ImageFile.ImageFile):
|
||||
format = "FITS"
|
||||
format_description = "FITS"
|
||||
|
||||
def _open(self) -> None:
|
||||
assert self.fp is not None
|
||||
|
||||
headers: dict[bytes, bytes] = {}
|
||||
header_in_progress = False
|
||||
decoder_name = ""
|
||||
while True:
|
||||
header = self.fp.read(80)
|
||||
if not header:
|
||||
msg = "Truncated FITS file"
|
||||
raise OSError(msg)
|
||||
keyword = header[:8].strip()
|
||||
if keyword in (b"SIMPLE", b"XTENSION"):
|
||||
header_in_progress = True
|
||||
elif headers and not header_in_progress:
|
||||
# This is now a data unit
|
||||
break
|
||||
elif keyword == b"END":
|
||||
# Seek to the end of the header unit
|
||||
self.fp.seek(math.ceil(self.fp.tell() / 2880) * 2880)
|
||||
if not decoder_name:
|
||||
decoder_name, offset, args = self._parse_headers(headers)
|
||||
|
||||
header_in_progress = False
|
||||
continue
|
||||
|
||||
if decoder_name:
|
||||
# Keep going to read past the headers
|
||||
continue
|
||||
|
||||
value = header[8:].split(b"/")[0].strip()
|
||||
if value.startswith(b"="):
|
||||
value = value[1:].strip()
|
||||
if not headers and (not _accept(keyword) or value != b"T"):
|
||||
msg = "Not a FITS file"
|
||||
raise SyntaxError(msg)
|
||||
headers[keyword] = value
|
||||
|
||||
if not decoder_name:
|
||||
msg = "No image data"
|
||||
raise ValueError(msg)
|
||||
|
||||
offset += self.fp.tell() - 80
|
||||
self.tile = [ImageFile._Tile(decoder_name, (0, 0) + self.size, offset, args)]
|
||||
|
||||
def _get_size(
|
||||
self, headers: dict[bytes, bytes], prefix: bytes
|
||||
) -> tuple[int, int] | None:
|
||||
naxis = int(headers[prefix + b"NAXIS"])
|
||||
if naxis == 0:
|
||||
return None
|
||||
|
||||
if naxis == 1:
|
||||
return 1, int(headers[prefix + b"NAXIS1"])
|
||||
else:
|
||||
return int(headers[prefix + b"NAXIS1"]), int(headers[prefix + b"NAXIS2"])
|
||||
|
||||
def _parse_headers(
|
||||
self, headers: dict[bytes, bytes]
|
||||
) -> tuple[str, int, tuple[str | int, ...]]:
|
||||
prefix = b""
|
||||
decoder_name = "raw"
|
||||
offset = 0
|
||||
if (
|
||||
headers.get(b"XTENSION") == b"'BINTABLE'"
|
||||
and headers.get(b"ZIMAGE") == b"T"
|
||||
and headers[b"ZCMPTYPE"] == b"'GZIP_1 '"
|
||||
):
|
||||
no_prefix_size = self._get_size(headers, prefix) or (0, 0)
|
||||
number_of_bits = int(headers[b"BITPIX"])
|
||||
offset = no_prefix_size[0] * no_prefix_size[1] * (number_of_bits // 8)
|
||||
|
||||
prefix = b"Z"
|
||||
decoder_name = "fits_gzip"
|
||||
|
||||
size = self._get_size(headers, prefix)
|
||||
if not size:
|
||||
return "", 0, ()
|
||||
|
||||
self._size = size
|
||||
|
||||
number_of_bits = int(headers[prefix + b"BITPIX"])
|
||||
if number_of_bits == 8:
|
||||
self._mode = "L"
|
||||
elif number_of_bits == 16:
|
||||
self._mode = "I;16"
|
||||
elif number_of_bits == 32:
|
||||
self._mode = "I"
|
||||
elif number_of_bits in (-32, -64):
|
||||
self._mode = "F"
|
||||
|
||||
args: tuple[str | int, ...]
|
||||
if decoder_name == "raw":
|
||||
args = (self.mode, 0, -1)
|
||||
else:
|
||||
args = (number_of_bits,)
|
||||
return decoder_name, offset, args
|
||||
|
||||
|
||||
class FitsGzipDecoder(ImageFile.PyDecoder):
|
||||
_pulls_fd = True
|
||||
|
||||
def decode(self, buffer: bytes | Image.SupportsArrayInterface) -> tuple[int, int]:
|
||||
assert self.fd is not None
|
||||
value = gzip.decompress(self.fd.read())
|
||||
|
||||
rows = []
|
||||
offset = 0
|
||||
number_of_bits = min(self.args[0] // 8, 4)
|
||||
for y in range(self.state.ysize):
|
||||
row = bytearray()
|
||||
for x in range(self.state.xsize):
|
||||
row += value[offset + (4 - number_of_bits) : offset + 4]
|
||||
offset += 4
|
||||
rows.append(row)
|
||||
self.set_as_raw(bytes([pixel for row in rows[::-1] for pixel in row]))
|
||||
return -1, 0
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Registry
|
||||
|
||||
Image.register_open(FitsImageFile.format, FitsImageFile, _accept)
|
||||
Image.register_decoder("fits_gzip", FitsGzipDecoder)
|
||||
|
||||
Image.register_extensions(FitsImageFile.format, [".fit", ".fits"])
|
||||
175
testclass/lib/python3.12/site-packages/PIL/FliImagePlugin.py
Normal file
175
testclass/lib/python3.12/site-packages/PIL/FliImagePlugin.py
Normal file
@ -0,0 +1,175 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# FLI/FLC file handling.
|
||||
#
|
||||
# History:
|
||||
# 95-09-01 fl Created
|
||||
# 97-01-03 fl Fixed parser, setup decoder tile
|
||||
# 98-07-15 fl Renamed offset attribute to avoid name clash
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997-98.
|
||||
# Copyright (c) Fredrik Lundh 1995-97.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import os
|
||||
|
||||
from . import Image, ImageFile, ImagePalette
|
||||
from ._binary import i16le as i16
|
||||
from ._binary import i32le as i32
|
||||
from ._binary import o8
|
||||
|
||||
#
|
||||
# decoder
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return (
|
||||
len(prefix) >= 6
|
||||
and i16(prefix, 4) in [0xAF11, 0xAF12]
|
||||
and i16(prefix, 14) in [0, 3] # flags
|
||||
)
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for the FLI/FLC animation format. Use the <b>seek</b>
|
||||
# method to load individual frames.
|
||||
|
||||
|
||||
class FliImageFile(ImageFile.ImageFile):
|
||||
format = "FLI"
|
||||
format_description = "Autodesk FLI/FLC Animation"
|
||||
_close_exclusive_fp_after_loading = False
|
||||
|
||||
def _open(self) -> None:
|
||||
# HEAD
|
||||
s = self.fp.read(128)
|
||||
if not (_accept(s) and s[20:22] == b"\x00\x00"):
|
||||
msg = "not an FLI/FLC file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
# frames
|
||||
self.n_frames = i16(s, 6)
|
||||
self.is_animated = self.n_frames > 1
|
||||
|
||||
# image characteristics
|
||||
self._mode = "P"
|
||||
self._size = i16(s, 8), i16(s, 10)
|
||||
|
||||
# animation speed
|
||||
duration = i32(s, 16)
|
||||
magic = i16(s, 4)
|
||||
if magic == 0xAF11:
|
||||
duration = (duration * 1000) // 70
|
||||
self.info["duration"] = duration
|
||||
|
||||
# look for palette
|
||||
palette = [(a, a, a) for a in range(256)]
|
||||
|
||||
s = self.fp.read(16)
|
||||
|
||||
self.__offset = 128
|
||||
|
||||
if i16(s, 4) == 0xF100:
|
||||
# prefix chunk; ignore it
|
||||
self.__offset = self.__offset + i32(s)
|
||||
self.fp.seek(self.__offset)
|
||||
s = self.fp.read(16)
|
||||
|
||||
if i16(s, 4) == 0xF1FA:
|
||||
# look for palette chunk
|
||||
number_of_subchunks = i16(s, 6)
|
||||
chunk_size: int | None = None
|
||||
for _ in range(number_of_subchunks):
|
||||
if chunk_size is not None:
|
||||
self.fp.seek(chunk_size - 6, os.SEEK_CUR)
|
||||
s = self.fp.read(6)
|
||||
chunk_type = i16(s, 4)
|
||||
if chunk_type in (4, 11):
|
||||
self._palette(palette, 2 if chunk_type == 11 else 0)
|
||||
break
|
||||
chunk_size = i32(s)
|
||||
if not chunk_size:
|
||||
break
|
||||
|
||||
self.palette = ImagePalette.raw(
|
||||
"RGB", b"".join(o8(r) + o8(g) + o8(b) for (r, g, b) in palette)
|
||||
)
|
||||
|
||||
# set things up to decode first frame
|
||||
self.__frame = -1
|
||||
self._fp = self.fp
|
||||
self.__rewind = self.fp.tell()
|
||||
self.seek(0)
|
||||
|
||||
def _palette(self, palette: list[tuple[int, int, int]], shift: int) -> None:
|
||||
# load palette
|
||||
|
||||
i = 0
|
||||
for e in range(i16(self.fp.read(2))):
|
||||
s = self.fp.read(2)
|
||||
i = i + s[0]
|
||||
n = s[1]
|
||||
if n == 0:
|
||||
n = 256
|
||||
s = self.fp.read(n * 3)
|
||||
for n in range(0, len(s), 3):
|
||||
r = s[n] << shift
|
||||
g = s[n + 1] << shift
|
||||
b = s[n + 2] << shift
|
||||
palette[i] = (r, g, b)
|
||||
i += 1
|
||||
|
||||
def seek(self, frame: int) -> None:
|
||||
if not self._seek_check(frame):
|
||||
return
|
||||
if frame < self.__frame:
|
||||
self._seek(0)
|
||||
|
||||
for f in range(self.__frame + 1, frame + 1):
|
||||
self._seek(f)
|
||||
|
||||
def _seek(self, frame: int) -> None:
|
||||
if frame == 0:
|
||||
self.__frame = -1
|
||||
self._fp.seek(self.__rewind)
|
||||
self.__offset = 128
|
||||
else:
|
||||
# ensure that the previous frame was loaded
|
||||
self.load()
|
||||
|
||||
if frame != self.__frame + 1:
|
||||
msg = f"cannot seek to frame {frame}"
|
||||
raise ValueError(msg)
|
||||
self.__frame = frame
|
||||
|
||||
# move to next frame
|
||||
self.fp = self._fp
|
||||
self.fp.seek(self.__offset)
|
||||
|
||||
s = self.fp.read(4)
|
||||
if not s:
|
||||
msg = "missing frame size"
|
||||
raise EOFError(msg)
|
||||
|
||||
framesize = i32(s)
|
||||
|
||||
self.decodermaxblock = framesize
|
||||
self.tile = [ImageFile._Tile("fli", (0, 0) + self.size, self.__offset)]
|
||||
|
||||
self.__offset += framesize
|
||||
|
||||
def tell(self) -> int:
|
||||
return self.__frame
|
||||
|
||||
|
||||
#
|
||||
# registry
|
||||
|
||||
Image.register_open(FliImageFile.format, FliImageFile, _accept)
|
||||
|
||||
Image.register_extensions(FliImageFile.format, [".fli", ".flc"])
|
||||
134
testclass/lib/python3.12/site-packages/PIL/FontFile.py
Normal file
134
testclass/lib/python3.12/site-packages/PIL/FontFile.py
Normal file
@ -0,0 +1,134 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# base class for raster font file parsers
|
||||
#
|
||||
# history:
|
||||
# 1997-06-05 fl created
|
||||
# 1997-08-19 fl restrict image width
|
||||
#
|
||||
# Copyright (c) 1997-1998 by Secret Labs AB
|
||||
# Copyright (c) 1997-1998 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import os
|
||||
from typing import BinaryIO
|
||||
|
||||
from . import Image, _binary
|
||||
|
||||
WIDTH = 800
|
||||
|
||||
|
||||
def puti16(
|
||||
fp: BinaryIO, values: tuple[int, int, int, int, int, int, int, int, int, int]
|
||||
) -> None:
|
||||
"""Write network order (big-endian) 16-bit sequence"""
|
||||
for v in values:
|
||||
if v < 0:
|
||||
v += 65536
|
||||
fp.write(_binary.o16be(v))
|
||||
|
||||
|
||||
class FontFile:
|
||||
"""Base class for raster font file handlers."""
|
||||
|
||||
bitmap: Image.Image | None = None
|
||||
|
||||
def __init__(self) -> None:
|
||||
self.info: dict[bytes, bytes | int] = {}
|
||||
self.glyph: list[
|
||||
tuple[
|
||||
tuple[int, int],
|
||||
tuple[int, int, int, int],
|
||||
tuple[int, int, int, int],
|
||||
Image.Image,
|
||||
]
|
||||
| None
|
||||
] = [None] * 256
|
||||
|
||||
def __getitem__(self, ix: int) -> (
|
||||
tuple[
|
||||
tuple[int, int],
|
||||
tuple[int, int, int, int],
|
||||
tuple[int, int, int, int],
|
||||
Image.Image,
|
||||
]
|
||||
| None
|
||||
):
|
||||
return self.glyph[ix]
|
||||
|
||||
def compile(self) -> None:
|
||||
"""Create metrics and bitmap"""
|
||||
|
||||
if self.bitmap:
|
||||
return
|
||||
|
||||
# create bitmap large enough to hold all data
|
||||
h = w = maxwidth = 0
|
||||
lines = 1
|
||||
for glyph in self.glyph:
|
||||
if glyph:
|
||||
d, dst, src, im = glyph
|
||||
h = max(h, src[3] - src[1])
|
||||
w = w + (src[2] - src[0])
|
||||
if w > WIDTH:
|
||||
lines += 1
|
||||
w = src[2] - src[0]
|
||||
maxwidth = max(maxwidth, w)
|
||||
|
||||
xsize = maxwidth
|
||||
ysize = lines * h
|
||||
|
||||
if xsize == 0 and ysize == 0:
|
||||
return
|
||||
|
||||
self.ysize = h
|
||||
|
||||
# paste glyphs into bitmap
|
||||
self.bitmap = Image.new("1", (xsize, ysize))
|
||||
self.metrics: list[
|
||||
tuple[tuple[int, int], tuple[int, int, int, int], tuple[int, int, int, int]]
|
||||
| None
|
||||
] = [None] * 256
|
||||
x = y = 0
|
||||
for i in range(256):
|
||||
glyph = self[i]
|
||||
if glyph:
|
||||
d, dst, src, im = glyph
|
||||
xx = src[2] - src[0]
|
||||
x0, y0 = x, y
|
||||
x = x + xx
|
||||
if x > WIDTH:
|
||||
x, y = 0, y + h
|
||||
x0, y0 = x, y
|
||||
x = xx
|
||||
s = src[0] + x0, src[1] + y0, src[2] + x0, src[3] + y0
|
||||
self.bitmap.paste(im.crop(src), s)
|
||||
self.metrics[i] = d, dst, s
|
||||
|
||||
def save(self, filename: str) -> None:
|
||||
"""Save font"""
|
||||
|
||||
self.compile()
|
||||
|
||||
# font data
|
||||
if not self.bitmap:
|
||||
msg = "No bitmap created"
|
||||
raise ValueError(msg)
|
||||
self.bitmap.save(os.path.splitext(filename)[0] + ".pbm", "PNG")
|
||||
|
||||
# font metrics
|
||||
with open(os.path.splitext(filename)[0] + ".pil", "wb") as fp:
|
||||
fp.write(b"PILfont\n")
|
||||
fp.write(f";;;;;;{self.ysize};\n".encode("ascii")) # HACK!!!
|
||||
fp.write(b"DATA\n")
|
||||
for id in range(256):
|
||||
m = self.metrics[id]
|
||||
if not m:
|
||||
puti16(fp, (0,) * 10)
|
||||
else:
|
||||
puti16(fp, m[0] + m[1] + m[2])
|
||||
257
testclass/lib/python3.12/site-packages/PIL/FpxImagePlugin.py
Normal file
257
testclass/lib/python3.12/site-packages/PIL/FpxImagePlugin.py
Normal file
@ -0,0 +1,257 @@
|
||||
#
|
||||
# THIS IS WORK IN PROGRESS
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# FlashPix support for PIL
|
||||
#
|
||||
# History:
|
||||
# 97-01-25 fl Created (reads uncompressed RGB images only)
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997.
|
||||
# Copyright (c) Fredrik Lundh 1997.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import olefile
|
||||
|
||||
from . import Image, ImageFile
|
||||
from ._binary import i32le as i32
|
||||
|
||||
# we map from colour field tuples to (mode, rawmode) descriptors
|
||||
MODES = {
|
||||
# opacity
|
||||
(0x00007FFE,): ("A", "L"),
|
||||
# monochrome
|
||||
(0x00010000,): ("L", "L"),
|
||||
(0x00018000, 0x00017FFE): ("RGBA", "LA"),
|
||||
# photo YCC
|
||||
(0x00020000, 0x00020001, 0x00020002): ("RGB", "YCC;P"),
|
||||
(0x00028000, 0x00028001, 0x00028002, 0x00027FFE): ("RGBA", "YCCA;P"),
|
||||
# standard RGB (NIFRGB)
|
||||
(0x00030000, 0x00030001, 0x00030002): ("RGB", "RGB"),
|
||||
(0x00038000, 0x00038001, 0x00038002, 0x00037FFE): ("RGBA", "RGBA"),
|
||||
}
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:8] == olefile.MAGIC
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for the FlashPix images.
|
||||
|
||||
|
||||
class FpxImageFile(ImageFile.ImageFile):
|
||||
format = "FPX"
|
||||
format_description = "FlashPix"
|
||||
|
||||
def _open(self) -> None:
|
||||
#
|
||||
# read the OLE directory and see if this is a likely
|
||||
# to be a FlashPix file
|
||||
|
||||
try:
|
||||
self.ole = olefile.OleFileIO(self.fp)
|
||||
except OSError as e:
|
||||
msg = "not an FPX file; invalid OLE file"
|
||||
raise SyntaxError(msg) from e
|
||||
|
||||
root = self.ole.root
|
||||
if not root or root.clsid != "56616700-C154-11CE-8553-00AA00A1F95B":
|
||||
msg = "not an FPX file; bad root CLSID"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
self._open_index(1)
|
||||
|
||||
def _open_index(self, index: int = 1) -> None:
|
||||
#
|
||||
# get the Image Contents Property Set
|
||||
|
||||
prop = self.ole.getproperties(
|
||||
[f"Data Object Store {index:06d}", "\005Image Contents"]
|
||||
)
|
||||
|
||||
# size (highest resolution)
|
||||
|
||||
assert isinstance(prop[0x1000002], int)
|
||||
assert isinstance(prop[0x1000003], int)
|
||||
self._size = prop[0x1000002], prop[0x1000003]
|
||||
|
||||
size = max(self.size)
|
||||
i = 1
|
||||
while size > 64:
|
||||
size = size // 2
|
||||
i += 1
|
||||
self.maxid = i - 1
|
||||
|
||||
# mode. instead of using a single field for this, flashpix
|
||||
# requires you to specify the mode for each channel in each
|
||||
# resolution subimage, and leaves it to the decoder to make
|
||||
# sure that they all match. for now, we'll cheat and assume
|
||||
# that this is always the case.
|
||||
|
||||
id = self.maxid << 16
|
||||
|
||||
s = prop[0x2000002 | id]
|
||||
|
||||
if not isinstance(s, bytes) or (bands := i32(s, 4)) > 4:
|
||||
msg = "Invalid number of bands"
|
||||
raise OSError(msg)
|
||||
|
||||
# note: for now, we ignore the "uncalibrated" flag
|
||||
colors = tuple(i32(s, 8 + i * 4) & 0x7FFFFFFF for i in range(bands))
|
||||
|
||||
self._mode, self.rawmode = MODES[colors]
|
||||
|
||||
# load JPEG tables, if any
|
||||
self.jpeg = {}
|
||||
for i in range(256):
|
||||
id = 0x3000001 | (i << 16)
|
||||
if id in prop:
|
||||
self.jpeg[i] = prop[id]
|
||||
|
||||
self._open_subimage(1, self.maxid)
|
||||
|
||||
def _open_subimage(self, index: int = 1, subimage: int = 0) -> None:
|
||||
#
|
||||
# setup tile descriptors for a given subimage
|
||||
|
||||
stream = [
|
||||
f"Data Object Store {index:06d}",
|
||||
f"Resolution {subimage:04d}",
|
||||
"Subimage 0000 Header",
|
||||
]
|
||||
|
||||
fp = self.ole.openstream(stream)
|
||||
|
||||
# skip prefix
|
||||
fp.read(28)
|
||||
|
||||
# header stream
|
||||
s = fp.read(36)
|
||||
|
||||
size = i32(s, 4), i32(s, 8)
|
||||
# tilecount = i32(s, 12)
|
||||
tilesize = i32(s, 16), i32(s, 20)
|
||||
# channels = i32(s, 24)
|
||||
offset = i32(s, 28)
|
||||
length = i32(s, 32)
|
||||
|
||||
if size != self.size:
|
||||
msg = "subimage mismatch"
|
||||
raise OSError(msg)
|
||||
|
||||
# get tile descriptors
|
||||
fp.seek(28 + offset)
|
||||
s = fp.read(i32(s, 12) * length)
|
||||
|
||||
x = y = 0
|
||||
xsize, ysize = size
|
||||
xtile, ytile = tilesize
|
||||
self.tile = []
|
||||
|
||||
for i in range(0, len(s), length):
|
||||
x1 = min(xsize, x + xtile)
|
||||
y1 = min(ysize, y + ytile)
|
||||
|
||||
compression = i32(s, i + 8)
|
||||
|
||||
if compression == 0:
|
||||
self.tile.append(
|
||||
ImageFile._Tile(
|
||||
"raw",
|
||||
(x, y, x1, y1),
|
||||
i32(s, i) + 28,
|
||||
self.rawmode,
|
||||
)
|
||||
)
|
||||
|
||||
elif compression == 1:
|
||||
# FIXME: the fill decoder is not implemented
|
||||
self.tile.append(
|
||||
ImageFile._Tile(
|
||||
"fill",
|
||||
(x, y, x1, y1),
|
||||
i32(s, i) + 28,
|
||||
(self.rawmode, s[12:16]),
|
||||
)
|
||||
)
|
||||
|
||||
elif compression == 2:
|
||||
internal_color_conversion = s[14]
|
||||
jpeg_tables = s[15]
|
||||
rawmode = self.rawmode
|
||||
|
||||
if internal_color_conversion:
|
||||
# The image is stored as usual (usually YCbCr).
|
||||
if rawmode == "RGBA":
|
||||
# For "RGBA", data is stored as YCbCrA based on
|
||||
# negative RGB. The following trick works around
|
||||
# this problem :
|
||||
jpegmode, rawmode = "YCbCrK", "CMYK"
|
||||
else:
|
||||
jpegmode = None # let the decoder decide
|
||||
|
||||
else:
|
||||
# The image is stored as defined by rawmode
|
||||
jpegmode = rawmode
|
||||
|
||||
self.tile.append(
|
||||
ImageFile._Tile(
|
||||
"jpeg",
|
||||
(x, y, x1, y1),
|
||||
i32(s, i) + 28,
|
||||
(rawmode, jpegmode),
|
||||
)
|
||||
)
|
||||
|
||||
# FIXME: jpeg tables are tile dependent; the prefix
|
||||
# data must be placed in the tile descriptor itself!
|
||||
|
||||
if jpeg_tables:
|
||||
self.tile_prefix = self.jpeg[jpeg_tables]
|
||||
|
||||
else:
|
||||
msg = "unknown/invalid compression"
|
||||
raise OSError(msg)
|
||||
|
||||
x = x + xtile
|
||||
if x >= xsize:
|
||||
x, y = 0, y + ytile
|
||||
if y >= ysize:
|
||||
break # isn't really required
|
||||
|
||||
self.stream = stream
|
||||
self._fp = self.fp
|
||||
self.fp = None
|
||||
|
||||
def load(self) -> Image.core.PixelAccess | None:
|
||||
if not self.fp:
|
||||
self.fp = self.ole.openstream(self.stream[:2] + ["Subimage 0000 Data"])
|
||||
|
||||
return ImageFile.ImageFile.load(self)
|
||||
|
||||
def close(self) -> None:
|
||||
self.ole.close()
|
||||
super().close()
|
||||
|
||||
def __exit__(self, *args: object) -> None:
|
||||
self.ole.close()
|
||||
super().__exit__()
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
|
||||
Image.register_open(FpxImageFile.format, FpxImageFile, _accept)
|
||||
|
||||
Image.register_extension(FpxImageFile.format, ".fpx")
|
||||
115
testclass/lib/python3.12/site-packages/PIL/FtexImagePlugin.py
Normal file
115
testclass/lib/python3.12/site-packages/PIL/FtexImagePlugin.py
Normal file
@ -0,0 +1,115 @@
|
||||
"""
|
||||
A Pillow loader for .ftc and .ftu files (FTEX)
|
||||
Jerome Leclanche <jerome@leclan.ch>
|
||||
|
||||
The contents of this file are hereby released in the public domain (CC0)
|
||||
Full text of the CC0 license:
|
||||
https://creativecommons.org/publicdomain/zero/1.0/
|
||||
|
||||
Independence War 2: Edge Of Chaos - Texture File Format - 16 October 2001
|
||||
|
||||
The textures used for 3D objects in Independence War 2: Edge Of Chaos are in a
|
||||
packed custom format called FTEX. This file format uses file extensions FTC
|
||||
and FTU.
|
||||
* FTC files are compressed textures (using standard texture compression).
|
||||
* FTU files are not compressed.
|
||||
Texture File Format
|
||||
The FTC and FTU texture files both use the same format. This
|
||||
has the following structure:
|
||||
{header}
|
||||
{format_directory}
|
||||
{data}
|
||||
Where:
|
||||
{header} = {
|
||||
u32:magic,
|
||||
u32:version,
|
||||
u32:width,
|
||||
u32:height,
|
||||
u32:mipmap_count,
|
||||
u32:format_count
|
||||
}
|
||||
|
||||
* The "magic" number is "FTEX".
|
||||
* "width" and "height" are the dimensions of the texture.
|
||||
* "mipmap_count" is the number of mipmaps in the texture.
|
||||
* "format_count" is the number of texture formats (different versions of the
|
||||
same texture) in this file.
|
||||
|
||||
{format_directory} = format_count * { u32:format, u32:where }
|
||||
|
||||
The format value is 0 for DXT1 compressed textures and 1 for 24-bit RGB
|
||||
uncompressed textures.
|
||||
The texture data for a format starts at the position "where" in the file.
|
||||
|
||||
Each set of texture data in the file has the following structure:
|
||||
{data} = format_count * { u32:mipmap_size, mipmap_size * { u8 } }
|
||||
* "mipmap_size" is the number of bytes in that mip level. For compressed
|
||||
textures this is the size of the texture data compressed with DXT1. For 24 bit
|
||||
uncompressed textures, this is 3 * width * height. Following this are the image
|
||||
bytes for that mipmap level.
|
||||
|
||||
Note: All data is stored in little-Endian (Intel) byte order.
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import struct
|
||||
from enum import IntEnum
|
||||
from io import BytesIO
|
||||
|
||||
from . import Image, ImageFile
|
||||
|
||||
MAGIC = b"FTEX"
|
||||
|
||||
|
||||
class Format(IntEnum):
|
||||
DXT1 = 0
|
||||
UNCOMPRESSED = 1
|
||||
|
||||
|
||||
class FtexImageFile(ImageFile.ImageFile):
|
||||
format = "FTEX"
|
||||
format_description = "Texture File Format (IW2:EOC)"
|
||||
|
||||
def _open(self) -> None:
|
||||
if not _accept(self.fp.read(4)):
|
||||
msg = "not an FTEX file"
|
||||
raise SyntaxError(msg)
|
||||
struct.unpack("<i", self.fp.read(4)) # version
|
||||
self._size = struct.unpack("<2i", self.fp.read(8))
|
||||
mipmap_count, format_count = struct.unpack("<2i", self.fp.read(8))
|
||||
|
||||
self._mode = "RGB"
|
||||
|
||||
# Only support single-format files.
|
||||
# I don't know of any multi-format file.
|
||||
assert format_count == 1
|
||||
|
||||
format, where = struct.unpack("<2i", self.fp.read(8))
|
||||
self.fp.seek(where)
|
||||
(mipmap_size,) = struct.unpack("<i", self.fp.read(4))
|
||||
|
||||
data = self.fp.read(mipmap_size)
|
||||
|
||||
if format == Format.DXT1:
|
||||
self._mode = "RGBA"
|
||||
self.tile = [ImageFile._Tile("bcn", (0, 0) + self.size, 0, (1,))]
|
||||
elif format == Format.UNCOMPRESSED:
|
||||
self.tile = [ImageFile._Tile("raw", (0, 0) + self.size, 0, "RGB")]
|
||||
else:
|
||||
msg = f"Invalid texture compression format: {repr(format)}"
|
||||
raise ValueError(msg)
|
||||
|
||||
self.fp.close()
|
||||
self.fp = BytesIO(data)
|
||||
|
||||
def load_seek(self, pos: int) -> None:
|
||||
pass
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == MAGIC
|
||||
|
||||
|
||||
Image.register_open(FtexImageFile.format, FtexImageFile, _accept)
|
||||
Image.register_extensions(FtexImageFile.format, [".ftc", ".ftu"])
|
||||
103
testclass/lib/python3.12/site-packages/PIL/GbrImagePlugin.py
Normal file
103
testclass/lib/python3.12/site-packages/PIL/GbrImagePlugin.py
Normal file
@ -0,0 +1,103 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
#
|
||||
# load a GIMP brush file
|
||||
#
|
||||
# History:
|
||||
# 96-03-14 fl Created
|
||||
# 16-01-08 es Version 2
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997.
|
||||
# Copyright (c) Fredrik Lundh 1996.
|
||||
# Copyright (c) Eric Soroos 2016.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
#
|
||||
# See https://github.com/GNOME/gimp/blob/mainline/devel-docs/gbr.txt for
|
||||
# format documentation.
|
||||
#
|
||||
# This code Interprets version 1 and 2 .gbr files.
|
||||
# Version 1 files are obsolete, and should not be used for new
|
||||
# brushes.
|
||||
# Version 2 files are saved by GIMP v2.8 (at least)
|
||||
# Version 3 files have a format specifier of 18 for 16bit floats in
|
||||
# the color depth field. This is currently unsupported by Pillow.
|
||||
from __future__ import annotations
|
||||
|
||||
from . import Image, ImageFile
|
||||
from ._binary import i32be as i32
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return len(prefix) >= 8 and i32(prefix, 0) >= 20 and i32(prefix, 4) in (1, 2)
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for the GIMP brush format.
|
||||
|
||||
|
||||
class GbrImageFile(ImageFile.ImageFile):
|
||||
format = "GBR"
|
||||
format_description = "GIMP brush file"
|
||||
|
||||
def _open(self) -> None:
|
||||
header_size = i32(self.fp.read(4))
|
||||
if header_size < 20:
|
||||
msg = "not a GIMP brush"
|
||||
raise SyntaxError(msg)
|
||||
version = i32(self.fp.read(4))
|
||||
if version not in (1, 2):
|
||||
msg = f"Unsupported GIMP brush version: {version}"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
width = i32(self.fp.read(4))
|
||||
height = i32(self.fp.read(4))
|
||||
color_depth = i32(self.fp.read(4))
|
||||
if width <= 0 or height <= 0:
|
||||
msg = "not a GIMP brush"
|
||||
raise SyntaxError(msg)
|
||||
if color_depth not in (1, 4):
|
||||
msg = f"Unsupported GIMP brush color depth: {color_depth}"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
if version == 1:
|
||||
comment_length = header_size - 20
|
||||
else:
|
||||
comment_length = header_size - 28
|
||||
magic_number = self.fp.read(4)
|
||||
if magic_number != b"GIMP":
|
||||
msg = "not a GIMP brush, bad magic number"
|
||||
raise SyntaxError(msg)
|
||||
self.info["spacing"] = i32(self.fp.read(4))
|
||||
|
||||
comment = self.fp.read(comment_length)[:-1]
|
||||
|
||||
if color_depth == 1:
|
||||
self._mode = "L"
|
||||
else:
|
||||
self._mode = "RGBA"
|
||||
|
||||
self._size = width, height
|
||||
|
||||
self.info["comment"] = comment
|
||||
|
||||
# Image might not be small
|
||||
Image._decompression_bomb_check(self.size)
|
||||
|
||||
# Data is an uncompressed block of w * h * bytes/pixel
|
||||
self._data_size = width * height * color_depth
|
||||
|
||||
def load(self) -> Image.core.PixelAccess | None:
|
||||
if self._im is None:
|
||||
self.im = Image.core.new(self.mode, self.size)
|
||||
self.frombytes(self.fp.read(self._data_size))
|
||||
return Image.Image.load(self)
|
||||
|
||||
|
||||
#
|
||||
# registry
|
||||
|
||||
|
||||
Image.register_open(GbrImageFile.format, GbrImageFile, _accept)
|
||||
Image.register_extension(GbrImageFile.format, ".gbr")
|
||||
102
testclass/lib/python3.12/site-packages/PIL/GdImageFile.py
Normal file
102
testclass/lib/python3.12/site-packages/PIL/GdImageFile.py
Normal file
@ -0,0 +1,102 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# GD file handling
|
||||
#
|
||||
# History:
|
||||
# 1996-04-12 fl Created
|
||||
#
|
||||
# Copyright (c) 1997 by Secret Labs AB.
|
||||
# Copyright (c) 1996 by Fredrik Lundh.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
|
||||
|
||||
"""
|
||||
.. note::
|
||||
This format cannot be automatically recognized, so the
|
||||
class is not registered for use with :py:func:`PIL.Image.open()`. To open a
|
||||
gd file, use the :py:func:`PIL.GdImageFile.open()` function instead.
|
||||
|
||||
.. warning::
|
||||
THE GD FORMAT IS NOT DESIGNED FOR DATA INTERCHANGE. This
|
||||
implementation is provided for convenience and demonstrational
|
||||
purposes only.
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import IO
|
||||
|
||||
from . import ImageFile, ImagePalette, UnidentifiedImageError
|
||||
from ._binary import i16be as i16
|
||||
from ._binary import i32be as i32
|
||||
from ._typing import StrOrBytesPath
|
||||
|
||||
|
||||
class GdImageFile(ImageFile.ImageFile):
|
||||
"""
|
||||
Image plugin for the GD uncompressed format. Note that this format
|
||||
is not supported by the standard :py:func:`PIL.Image.open()` function. To use
|
||||
this plugin, you have to import the :py:mod:`PIL.GdImageFile` module and
|
||||
use the :py:func:`PIL.GdImageFile.open()` function.
|
||||
"""
|
||||
|
||||
format = "GD"
|
||||
format_description = "GD uncompressed images"
|
||||
|
||||
def _open(self) -> None:
|
||||
# Header
|
||||
assert self.fp is not None
|
||||
|
||||
s = self.fp.read(1037)
|
||||
|
||||
if i16(s) not in [65534, 65535]:
|
||||
msg = "Not a valid GD 2.x .gd file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
self._mode = "L" # FIXME: "P"
|
||||
self._size = i16(s, 2), i16(s, 4)
|
||||
|
||||
true_color = s[6]
|
||||
true_color_offset = 2 if true_color else 0
|
||||
|
||||
# transparency index
|
||||
tindex = i32(s, 7 + true_color_offset)
|
||||
if tindex < 256:
|
||||
self.info["transparency"] = tindex
|
||||
|
||||
self.palette = ImagePalette.raw(
|
||||
"XBGR", s[7 + true_color_offset + 4 : 7 + true_color_offset + 4 + 256 * 4]
|
||||
)
|
||||
|
||||
self.tile = [
|
||||
ImageFile._Tile(
|
||||
"raw",
|
||||
(0, 0) + self.size,
|
||||
7 + true_color_offset + 4 + 256 * 4,
|
||||
"L",
|
||||
)
|
||||
]
|
||||
|
||||
|
||||
def open(fp: StrOrBytesPath | IO[bytes], mode: str = "r") -> GdImageFile:
|
||||
"""
|
||||
Load texture from a GD image file.
|
||||
|
||||
:param fp: GD file name, or an opened file handle.
|
||||
:param mode: Optional mode. In this version, if the mode argument
|
||||
is given, it must be "r".
|
||||
:returns: An image instance.
|
||||
:raises OSError: If the image could not be read.
|
||||
"""
|
||||
if mode != "r":
|
||||
msg = "bad mode"
|
||||
raise ValueError(msg)
|
||||
|
||||
try:
|
||||
return GdImageFile(fp)
|
||||
except SyntaxError as e:
|
||||
msg = "cannot identify this image file"
|
||||
raise UnidentifiedImageError(msg) from e
|
||||
1197
testclass/lib/python3.12/site-packages/PIL/GifImagePlugin.py
Normal file
1197
testclass/lib/python3.12/site-packages/PIL/GifImagePlugin.py
Normal file
File diff suppressed because it is too large
Load Diff
149
testclass/lib/python3.12/site-packages/PIL/GimpGradientFile.py
Normal file
149
testclass/lib/python3.12/site-packages/PIL/GimpGradientFile.py
Normal file
@ -0,0 +1,149 @@
|
||||
#
|
||||
# Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# stuff to read (and render) GIMP gradient files
|
||||
#
|
||||
# History:
|
||||
# 97-08-23 fl Created
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997.
|
||||
# Copyright (c) Fredrik Lundh 1997.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
|
||||
"""
|
||||
Stuff to translate curve segments to palette values (derived from
|
||||
the corresponding code in GIMP, written by Federico Mena Quintero.
|
||||
See the GIMP distribution for more information.)
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
from math import log, pi, sin, sqrt
|
||||
from typing import IO, Callable
|
||||
|
||||
from ._binary import o8
|
||||
|
||||
EPSILON = 1e-10
|
||||
"""""" # Enable auto-doc for data member
|
||||
|
||||
|
||||
def linear(middle: float, pos: float) -> float:
|
||||
if pos <= middle:
|
||||
if middle < EPSILON:
|
||||
return 0.0
|
||||
else:
|
||||
return 0.5 * pos / middle
|
||||
else:
|
||||
pos = pos - middle
|
||||
middle = 1.0 - middle
|
||||
if middle < EPSILON:
|
||||
return 1.0
|
||||
else:
|
||||
return 0.5 + 0.5 * pos / middle
|
||||
|
||||
|
||||
def curved(middle: float, pos: float) -> float:
|
||||
return pos ** (log(0.5) / log(max(middle, EPSILON)))
|
||||
|
||||
|
||||
def sine(middle: float, pos: float) -> float:
|
||||
return (sin((-pi / 2.0) + pi * linear(middle, pos)) + 1.0) / 2.0
|
||||
|
||||
|
||||
def sphere_increasing(middle: float, pos: float) -> float:
|
||||
return sqrt(1.0 - (linear(middle, pos) - 1.0) ** 2)
|
||||
|
||||
|
||||
def sphere_decreasing(middle: float, pos: float) -> float:
|
||||
return 1.0 - sqrt(1.0 - linear(middle, pos) ** 2)
|
||||
|
||||
|
||||
SEGMENTS = [linear, curved, sine, sphere_increasing, sphere_decreasing]
|
||||
"""""" # Enable auto-doc for data member
|
||||
|
||||
|
||||
class GradientFile:
|
||||
gradient: (
|
||||
list[
|
||||
tuple[
|
||||
float,
|
||||
float,
|
||||
float,
|
||||
list[float],
|
||||
list[float],
|
||||
Callable[[float, float], float],
|
||||
]
|
||||
]
|
||||
| None
|
||||
) = None
|
||||
|
||||
def getpalette(self, entries: int = 256) -> tuple[bytes, str]:
|
||||
assert self.gradient is not None
|
||||
palette = []
|
||||
|
||||
ix = 0
|
||||
x0, x1, xm, rgb0, rgb1, segment = self.gradient[ix]
|
||||
|
||||
for i in range(entries):
|
||||
x = i / (entries - 1)
|
||||
|
||||
while x1 < x:
|
||||
ix += 1
|
||||
x0, x1, xm, rgb0, rgb1, segment = self.gradient[ix]
|
||||
|
||||
w = x1 - x0
|
||||
|
||||
if w < EPSILON:
|
||||
scale = segment(0.5, 0.5)
|
||||
else:
|
||||
scale = segment((xm - x0) / w, (x - x0) / w)
|
||||
|
||||
# expand to RGBA
|
||||
r = o8(int(255 * ((rgb1[0] - rgb0[0]) * scale + rgb0[0]) + 0.5))
|
||||
g = o8(int(255 * ((rgb1[1] - rgb0[1]) * scale + rgb0[1]) + 0.5))
|
||||
b = o8(int(255 * ((rgb1[2] - rgb0[2]) * scale + rgb0[2]) + 0.5))
|
||||
a = o8(int(255 * ((rgb1[3] - rgb0[3]) * scale + rgb0[3]) + 0.5))
|
||||
|
||||
# add to palette
|
||||
palette.append(r + g + b + a)
|
||||
|
||||
return b"".join(palette), "RGBA"
|
||||
|
||||
|
||||
class GimpGradientFile(GradientFile):
|
||||
"""File handler for GIMP's gradient format."""
|
||||
|
||||
def __init__(self, fp: IO[bytes]) -> None:
|
||||
if fp.readline()[:13] != b"GIMP Gradient":
|
||||
msg = "not a GIMP gradient file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
line = fp.readline()
|
||||
|
||||
# GIMP 1.2 gradient files don't contain a name, but GIMP 1.3 files do
|
||||
if line.startswith(b"Name: "):
|
||||
line = fp.readline().strip()
|
||||
|
||||
count = int(line)
|
||||
|
||||
self.gradient = []
|
||||
|
||||
for i in range(count):
|
||||
s = fp.readline().split()
|
||||
w = [float(x) for x in s[:11]]
|
||||
|
||||
x0, x1 = w[0], w[2]
|
||||
xm = w[1]
|
||||
rgb0 = w[3:7]
|
||||
rgb1 = w[7:11]
|
||||
|
||||
segment = SEGMENTS[int(s[11])]
|
||||
cspace = int(s[12])
|
||||
|
||||
if cspace != 0:
|
||||
msg = "cannot handle HSV colour space"
|
||||
raise OSError(msg)
|
||||
|
||||
self.gradient.append((x0, x1, xm, rgb0, rgb1, segment))
|
||||
@ -0,0 +1,58 @@
|
||||
#
|
||||
# Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# stuff to read GIMP palette files
|
||||
#
|
||||
# History:
|
||||
# 1997-08-23 fl Created
|
||||
# 2004-09-07 fl Support GIMP 2.0 palette files.
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997-2004. All rights reserved.
|
||||
# Copyright (c) Fredrik Lundh 1997-2004.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import re
|
||||
from typing import IO
|
||||
|
||||
from ._binary import o8
|
||||
|
||||
|
||||
class GimpPaletteFile:
|
||||
"""File handler for GIMP's palette format."""
|
||||
|
||||
rawmode = "RGB"
|
||||
|
||||
def __init__(self, fp: IO[bytes]) -> None:
|
||||
palette = [o8(i) * 3 for i in range(256)]
|
||||
|
||||
if fp.readline()[:12] != b"GIMP Palette":
|
||||
msg = "not a GIMP palette file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
for i in range(256):
|
||||
s = fp.readline()
|
||||
if not s:
|
||||
break
|
||||
|
||||
# skip fields and comment lines
|
||||
if re.match(rb"\w+:|#", s):
|
||||
continue
|
||||
if len(s) > 100:
|
||||
msg = "bad palette file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
v = tuple(map(int, s.split()[:3]))
|
||||
if len(v) != 3:
|
||||
msg = "bad palette entry"
|
||||
raise ValueError(msg)
|
||||
|
||||
palette[i] = o8(v[0]) + o8(v[1]) + o8(v[2])
|
||||
|
||||
self.palette = b"".join(palette)
|
||||
|
||||
def getpalette(self) -> tuple[bytes, str]:
|
||||
return self.palette, self.rawmode
|
||||
@ -0,0 +1,76 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# GRIB stub adapter
|
||||
#
|
||||
# Copyright (c) 1996-2003 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import IO
|
||||
|
||||
from . import Image, ImageFile
|
||||
|
||||
_handler = None
|
||||
|
||||
|
||||
def register_handler(handler: ImageFile.StubHandler | None) -> None:
|
||||
"""
|
||||
Install application-specific GRIB image handler.
|
||||
|
||||
:param handler: Handler object.
|
||||
"""
|
||||
global _handler
|
||||
_handler = handler
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Image adapter
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == b"GRIB" and prefix[7] == 1
|
||||
|
||||
|
||||
class GribStubImageFile(ImageFile.StubImageFile):
|
||||
format = "GRIB"
|
||||
format_description = "GRIB"
|
||||
|
||||
def _open(self) -> None:
|
||||
offset = self.fp.tell()
|
||||
|
||||
if not _accept(self.fp.read(8)):
|
||||
msg = "Not a GRIB file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
self.fp.seek(offset)
|
||||
|
||||
# make something up
|
||||
self._mode = "F"
|
||||
self._size = 1, 1
|
||||
|
||||
loader = self._load()
|
||||
if loader:
|
||||
loader.open(self)
|
||||
|
||||
def _load(self) -> ImageFile.StubHandler | None:
|
||||
return _handler
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
if _handler is None or not hasattr(_handler, "save"):
|
||||
msg = "GRIB save handler not installed"
|
||||
raise OSError(msg)
|
||||
_handler.save(im, fp, filename)
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Registry
|
||||
|
||||
Image.register_open(GribStubImageFile.format, GribStubImageFile, _accept)
|
||||
Image.register_save(GribStubImageFile.format, _save)
|
||||
|
||||
Image.register_extension(GribStubImageFile.format, ".grib")
|
||||
@ -0,0 +1,76 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# HDF5 stub adapter
|
||||
#
|
||||
# Copyright (c) 2000-2003 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import IO
|
||||
|
||||
from . import Image, ImageFile
|
||||
|
||||
_handler = None
|
||||
|
||||
|
||||
def register_handler(handler: ImageFile.StubHandler | None) -> None:
|
||||
"""
|
||||
Install application-specific HDF5 image handler.
|
||||
|
||||
:param handler: Handler object.
|
||||
"""
|
||||
global _handler
|
||||
_handler = handler
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Image adapter
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:8] == b"\x89HDF\r\n\x1a\n"
|
||||
|
||||
|
||||
class HDF5StubImageFile(ImageFile.StubImageFile):
|
||||
format = "HDF5"
|
||||
format_description = "HDF5"
|
||||
|
||||
def _open(self) -> None:
|
||||
offset = self.fp.tell()
|
||||
|
||||
if not _accept(self.fp.read(8)):
|
||||
msg = "Not an HDF file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
self.fp.seek(offset)
|
||||
|
||||
# make something up
|
||||
self._mode = "F"
|
||||
self._size = 1, 1
|
||||
|
||||
loader = self._load()
|
||||
if loader:
|
||||
loader.open(self)
|
||||
|
||||
def _load(self) -> ImageFile.StubHandler | None:
|
||||
return _handler
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
if _handler is None or not hasattr(_handler, "save"):
|
||||
msg = "HDF5 save handler not installed"
|
||||
raise OSError(msg)
|
||||
_handler.save(im, fp, filename)
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Registry
|
||||
|
||||
Image.register_open(HDF5StubImageFile.format, HDF5StubImageFile, _accept)
|
||||
Image.register_save(HDF5StubImageFile.format, _save)
|
||||
|
||||
Image.register_extensions(HDF5StubImageFile.format, [".h5", ".hdf"])
|
||||
412
testclass/lib/python3.12/site-packages/PIL/IcnsImagePlugin.py
Normal file
412
testclass/lib/python3.12/site-packages/PIL/IcnsImagePlugin.py
Normal file
@ -0,0 +1,412 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# macOS icns file decoder, based on icns.py by Bob Ippolito.
|
||||
#
|
||||
# history:
|
||||
# 2004-10-09 fl Turned into a PIL plugin; removed 2.3 dependencies.
|
||||
# 2020-04-04 Allow saving on all operating systems.
|
||||
#
|
||||
# Copyright (c) 2004 by Bob Ippolito.
|
||||
# Copyright (c) 2004 by Secret Labs.
|
||||
# Copyright (c) 2004 by Fredrik Lundh.
|
||||
# Copyright (c) 2014 by Alastair Houghton.
|
||||
# Copyright (c) 2020 by Pan Jing.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import io
|
||||
import os
|
||||
import struct
|
||||
import sys
|
||||
from typing import IO
|
||||
|
||||
from . import Image, ImageFile, PngImagePlugin, features
|
||||
from ._deprecate import deprecate
|
||||
|
||||
enable_jpeg2k = features.check_codec("jpg_2000")
|
||||
if enable_jpeg2k:
|
||||
from . import Jpeg2KImagePlugin
|
||||
|
||||
MAGIC = b"icns"
|
||||
HEADERSIZE = 8
|
||||
|
||||
|
||||
def nextheader(fobj: IO[bytes]) -> tuple[bytes, int]:
|
||||
return struct.unpack(">4sI", fobj.read(HEADERSIZE))
|
||||
|
||||
|
||||
def read_32t(
|
||||
fobj: IO[bytes], start_length: tuple[int, int], size: tuple[int, int, int]
|
||||
) -> dict[str, Image.Image]:
|
||||
# The 128x128 icon seems to have an extra header for some reason.
|
||||
(start, length) = start_length
|
||||
fobj.seek(start)
|
||||
sig = fobj.read(4)
|
||||
if sig != b"\x00\x00\x00\x00":
|
||||
msg = "Unknown signature, expecting 0x00000000"
|
||||
raise SyntaxError(msg)
|
||||
return read_32(fobj, (start + 4, length - 4), size)
|
||||
|
||||
|
||||
def read_32(
|
||||
fobj: IO[bytes], start_length: tuple[int, int], size: tuple[int, int, int]
|
||||
) -> dict[str, Image.Image]:
|
||||
"""
|
||||
Read a 32bit RGB icon resource. Seems to be either uncompressed or
|
||||
an RLE packbits-like scheme.
|
||||
"""
|
||||
(start, length) = start_length
|
||||
fobj.seek(start)
|
||||
pixel_size = (size[0] * size[2], size[1] * size[2])
|
||||
sizesq = pixel_size[0] * pixel_size[1]
|
||||
if length == sizesq * 3:
|
||||
# uncompressed ("RGBRGBGB")
|
||||
indata = fobj.read(length)
|
||||
im = Image.frombuffer("RGB", pixel_size, indata, "raw", "RGB", 0, 1)
|
||||
else:
|
||||
# decode image
|
||||
im = Image.new("RGB", pixel_size, None)
|
||||
for band_ix in range(3):
|
||||
data = []
|
||||
bytesleft = sizesq
|
||||
while bytesleft > 0:
|
||||
byte = fobj.read(1)
|
||||
if not byte:
|
||||
break
|
||||
byte_int = byte[0]
|
||||
if byte_int & 0x80:
|
||||
blocksize = byte_int - 125
|
||||
byte = fobj.read(1)
|
||||
for i in range(blocksize):
|
||||
data.append(byte)
|
||||
else:
|
||||
blocksize = byte_int + 1
|
||||
data.append(fobj.read(blocksize))
|
||||
bytesleft -= blocksize
|
||||
if bytesleft <= 0:
|
||||
break
|
||||
if bytesleft != 0:
|
||||
msg = f"Error reading channel [{repr(bytesleft)} left]"
|
||||
raise SyntaxError(msg)
|
||||
band = Image.frombuffer("L", pixel_size, b"".join(data), "raw", "L", 0, 1)
|
||||
im.im.putband(band.im, band_ix)
|
||||
return {"RGB": im}
|
||||
|
||||
|
||||
def read_mk(
|
||||
fobj: IO[bytes], start_length: tuple[int, int], size: tuple[int, int, int]
|
||||
) -> dict[str, Image.Image]:
|
||||
# Alpha masks seem to be uncompressed
|
||||
start = start_length[0]
|
||||
fobj.seek(start)
|
||||
pixel_size = (size[0] * size[2], size[1] * size[2])
|
||||
sizesq = pixel_size[0] * pixel_size[1]
|
||||
band = Image.frombuffer("L", pixel_size, fobj.read(sizesq), "raw", "L", 0, 1)
|
||||
return {"A": band}
|
||||
|
||||
|
||||
def read_png_or_jpeg2000(
|
||||
fobj: IO[bytes], start_length: tuple[int, int], size: tuple[int, int, int]
|
||||
) -> dict[str, Image.Image]:
|
||||
(start, length) = start_length
|
||||
fobj.seek(start)
|
||||
sig = fobj.read(12)
|
||||
|
||||
im: Image.Image
|
||||
if sig[:8] == b"\x89PNG\x0d\x0a\x1a\x0a":
|
||||
fobj.seek(start)
|
||||
im = PngImagePlugin.PngImageFile(fobj)
|
||||
Image._decompression_bomb_check(im.size)
|
||||
return {"RGBA": im}
|
||||
elif (
|
||||
sig[:4] == b"\xff\x4f\xff\x51"
|
||||
or sig[:4] == b"\x0d\x0a\x87\x0a"
|
||||
or sig == b"\x00\x00\x00\x0cjP \x0d\x0a\x87\x0a"
|
||||
):
|
||||
if not enable_jpeg2k:
|
||||
msg = (
|
||||
"Unsupported icon subimage format (rebuild PIL "
|
||||
"with JPEG 2000 support to fix this)"
|
||||
)
|
||||
raise ValueError(msg)
|
||||
# j2k, jpc or j2c
|
||||
fobj.seek(start)
|
||||
jp2kstream = fobj.read(length)
|
||||
f = io.BytesIO(jp2kstream)
|
||||
im = Jpeg2KImagePlugin.Jpeg2KImageFile(f)
|
||||
Image._decompression_bomb_check(im.size)
|
||||
if im.mode != "RGBA":
|
||||
im = im.convert("RGBA")
|
||||
return {"RGBA": im}
|
||||
else:
|
||||
msg = "Unsupported icon subimage format"
|
||||
raise ValueError(msg)
|
||||
|
||||
|
||||
class IcnsFile:
|
||||
SIZES = {
|
||||
(512, 512, 2): [(b"ic10", read_png_or_jpeg2000)],
|
||||
(512, 512, 1): [(b"ic09", read_png_or_jpeg2000)],
|
||||
(256, 256, 2): [(b"ic14", read_png_or_jpeg2000)],
|
||||
(256, 256, 1): [(b"ic08", read_png_or_jpeg2000)],
|
||||
(128, 128, 2): [(b"ic13", read_png_or_jpeg2000)],
|
||||
(128, 128, 1): [
|
||||
(b"ic07", read_png_or_jpeg2000),
|
||||
(b"it32", read_32t),
|
||||
(b"t8mk", read_mk),
|
||||
],
|
||||
(64, 64, 1): [(b"icp6", read_png_or_jpeg2000)],
|
||||
(32, 32, 2): [(b"ic12", read_png_or_jpeg2000)],
|
||||
(48, 48, 1): [(b"ih32", read_32), (b"h8mk", read_mk)],
|
||||
(32, 32, 1): [
|
||||
(b"icp5", read_png_or_jpeg2000),
|
||||
(b"il32", read_32),
|
||||
(b"l8mk", read_mk),
|
||||
],
|
||||
(16, 16, 2): [(b"ic11", read_png_or_jpeg2000)],
|
||||
(16, 16, 1): [
|
||||
(b"icp4", read_png_or_jpeg2000),
|
||||
(b"is32", read_32),
|
||||
(b"s8mk", read_mk),
|
||||
],
|
||||
}
|
||||
|
||||
def __init__(self, fobj: IO[bytes]) -> None:
|
||||
"""
|
||||
fobj is a file-like object as an icns resource
|
||||
"""
|
||||
# signature : (start, length)
|
||||
self.dct = {}
|
||||
self.fobj = fobj
|
||||
sig, filesize = nextheader(fobj)
|
||||
if not _accept(sig):
|
||||
msg = "not an icns file"
|
||||
raise SyntaxError(msg)
|
||||
i = HEADERSIZE
|
||||
while i < filesize:
|
||||
sig, blocksize = nextheader(fobj)
|
||||
if blocksize <= 0:
|
||||
msg = "invalid block header"
|
||||
raise SyntaxError(msg)
|
||||
i += HEADERSIZE
|
||||
blocksize -= HEADERSIZE
|
||||
self.dct[sig] = (i, blocksize)
|
||||
fobj.seek(blocksize, io.SEEK_CUR)
|
||||
i += blocksize
|
||||
|
||||
def itersizes(self) -> list[tuple[int, int, int]]:
|
||||
sizes = []
|
||||
for size, fmts in self.SIZES.items():
|
||||
for fmt, reader in fmts:
|
||||
if fmt in self.dct:
|
||||
sizes.append(size)
|
||||
break
|
||||
return sizes
|
||||
|
||||
def bestsize(self) -> tuple[int, int, int]:
|
||||
sizes = self.itersizes()
|
||||
if not sizes:
|
||||
msg = "No 32bit icon resources found"
|
||||
raise SyntaxError(msg)
|
||||
return max(sizes)
|
||||
|
||||
def dataforsize(self, size: tuple[int, int, int]) -> dict[str, Image.Image]:
|
||||
"""
|
||||
Get an icon resource as {channel: array}. Note that
|
||||
the arrays are bottom-up like windows bitmaps and will likely
|
||||
need to be flipped or transposed in some way.
|
||||
"""
|
||||
dct = {}
|
||||
for code, reader in self.SIZES[size]:
|
||||
desc = self.dct.get(code)
|
||||
if desc is not None:
|
||||
dct.update(reader(self.fobj, desc, size))
|
||||
return dct
|
||||
|
||||
def getimage(
|
||||
self, size: tuple[int, int] | tuple[int, int, int] | None = None
|
||||
) -> Image.Image:
|
||||
if size is None:
|
||||
size = self.bestsize()
|
||||
elif len(size) == 2:
|
||||
size = (size[0], size[1], 1)
|
||||
channels = self.dataforsize(size)
|
||||
|
||||
im = channels.get("RGBA")
|
||||
if im:
|
||||
return im
|
||||
|
||||
im = channels["RGB"].copy()
|
||||
try:
|
||||
im.putalpha(channels["A"])
|
||||
except KeyError:
|
||||
pass
|
||||
return im
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for Mac OS icons.
|
||||
|
||||
|
||||
class IcnsImageFile(ImageFile.ImageFile):
|
||||
"""
|
||||
PIL image support for Mac OS .icns files.
|
||||
Chooses the best resolution, but will possibly load
|
||||
a different size image if you mutate the size attribute
|
||||
before calling 'load'.
|
||||
|
||||
The info dictionary has a key 'sizes' that is a list
|
||||
of sizes that the icns file has.
|
||||
"""
|
||||
|
||||
format = "ICNS"
|
||||
format_description = "Mac OS icns resource"
|
||||
|
||||
def _open(self) -> None:
|
||||
self.icns = IcnsFile(self.fp)
|
||||
self._mode = "RGBA"
|
||||
self.info["sizes"] = self.icns.itersizes()
|
||||
self.best_size = self.icns.bestsize()
|
||||
self.size = (
|
||||
self.best_size[0] * self.best_size[2],
|
||||
self.best_size[1] * self.best_size[2],
|
||||
)
|
||||
|
||||
@property # type: ignore[override]
|
||||
def size(self) -> tuple[int, int] | tuple[int, int, int]:
|
||||
return self._size
|
||||
|
||||
@size.setter
|
||||
def size(self, value: tuple[int, int] | tuple[int, int, int]) -> None:
|
||||
if len(value) == 3:
|
||||
deprecate("Setting size to (width, height, scale)", 12, "load(scale)")
|
||||
if value in self.info["sizes"]:
|
||||
self._size = value # type: ignore[assignment]
|
||||
return
|
||||
else:
|
||||
# Check that a matching size exists,
|
||||
# or that there is a scale that would create a size that matches
|
||||
for size in self.info["sizes"]:
|
||||
simple_size = size[0] * size[2], size[1] * size[2]
|
||||
scale = simple_size[0] // value[0]
|
||||
if simple_size[1] / value[1] == scale:
|
||||
self._size = value
|
||||
return
|
||||
msg = "This is not one of the allowed sizes of this image"
|
||||
raise ValueError(msg)
|
||||
|
||||
def load(self, scale: int | None = None) -> Image.core.PixelAccess | None:
|
||||
if scale is not None or len(self.size) == 3:
|
||||
if scale is None and len(self.size) == 3:
|
||||
scale = self.size[2]
|
||||
assert scale is not None
|
||||
width, height = self.size[:2]
|
||||
self.size = width * scale, height * scale
|
||||
self.best_size = width, height, scale
|
||||
|
||||
px = Image.Image.load(self)
|
||||
if self._im is not None and self.im.size == self.size:
|
||||
# Already loaded
|
||||
return px
|
||||
self.load_prepare()
|
||||
# This is likely NOT the best way to do it, but whatever.
|
||||
im = self.icns.getimage(self.best_size)
|
||||
|
||||
# If this is a PNG or JPEG 2000, it won't be loaded yet
|
||||
px = im.load()
|
||||
|
||||
self.im = im.im
|
||||
self._mode = im.mode
|
||||
self.size = im.size
|
||||
|
||||
return px
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
"""
|
||||
Saves the image as a series of PNG files,
|
||||
that are then combined into a .icns file.
|
||||
"""
|
||||
if hasattr(fp, "flush"):
|
||||
fp.flush()
|
||||
|
||||
sizes = {
|
||||
b"ic07": 128,
|
||||
b"ic08": 256,
|
||||
b"ic09": 512,
|
||||
b"ic10": 1024,
|
||||
b"ic11": 32,
|
||||
b"ic12": 64,
|
||||
b"ic13": 256,
|
||||
b"ic14": 512,
|
||||
}
|
||||
provided_images = {im.width: im for im in im.encoderinfo.get("append_images", [])}
|
||||
size_streams = {}
|
||||
for size in set(sizes.values()):
|
||||
image = (
|
||||
provided_images[size]
|
||||
if size in provided_images
|
||||
else im.resize((size, size))
|
||||
)
|
||||
|
||||
temp = io.BytesIO()
|
||||
image.save(temp, "png")
|
||||
size_streams[size] = temp.getvalue()
|
||||
|
||||
entries = []
|
||||
for type, size in sizes.items():
|
||||
stream = size_streams[size]
|
||||
entries.append((type, HEADERSIZE + len(stream), stream))
|
||||
|
||||
# Header
|
||||
fp.write(MAGIC)
|
||||
file_length = HEADERSIZE # Header
|
||||
file_length += HEADERSIZE + 8 * len(entries) # TOC
|
||||
file_length += sum(entry[1] for entry in entries)
|
||||
fp.write(struct.pack(">i", file_length))
|
||||
|
||||
# TOC
|
||||
fp.write(b"TOC ")
|
||||
fp.write(struct.pack(">i", HEADERSIZE + len(entries) * HEADERSIZE))
|
||||
for entry in entries:
|
||||
fp.write(entry[0])
|
||||
fp.write(struct.pack(">i", entry[1]))
|
||||
|
||||
# Data
|
||||
for entry in entries:
|
||||
fp.write(entry[0])
|
||||
fp.write(struct.pack(">i", entry[1]))
|
||||
fp.write(entry[2])
|
||||
|
||||
if hasattr(fp, "flush"):
|
||||
fp.flush()
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == MAGIC
|
||||
|
||||
|
||||
Image.register_open(IcnsImageFile.format, IcnsImageFile, _accept)
|
||||
Image.register_extension(IcnsImageFile.format, ".icns")
|
||||
|
||||
Image.register_save(IcnsImageFile.format, _save)
|
||||
Image.register_mime(IcnsImageFile.format, "image/icns")
|
||||
|
||||
if __name__ == "__main__":
|
||||
if len(sys.argv) < 2:
|
||||
print("Syntax: python3 IcnsImagePlugin.py [file]")
|
||||
sys.exit()
|
||||
|
||||
with open(sys.argv[1], "rb") as fp:
|
||||
imf = IcnsImageFile(fp)
|
||||
for size in imf.info["sizes"]:
|
||||
width, height, scale = imf.size = size
|
||||
imf.save(f"out-{width}-{height}-{scale}.png")
|
||||
with Image.open(sys.argv[1]) as im:
|
||||
im.save("out.png")
|
||||
if sys.platform == "windows":
|
||||
os.startfile("out.png")
|
||||
381
testclass/lib/python3.12/site-packages/PIL/IcoImagePlugin.py
Normal file
381
testclass/lib/python3.12/site-packages/PIL/IcoImagePlugin.py
Normal file
@ -0,0 +1,381 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# Windows Icon support for PIL
|
||||
#
|
||||
# History:
|
||||
# 96-05-27 fl Created
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997.
|
||||
# Copyright (c) Fredrik Lundh 1996.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
|
||||
# This plugin is a refactored version of Win32IconImagePlugin by Bryan Davis
|
||||
# <casadebender@gmail.com>.
|
||||
# https://code.google.com/archive/p/casadebender/wikis/Win32IconImagePlugin.wiki
|
||||
#
|
||||
# Icon format references:
|
||||
# * https://en.wikipedia.org/wiki/ICO_(file_format)
|
||||
# * https://msdn.microsoft.com/en-us/library/ms997538.aspx
|
||||
from __future__ import annotations
|
||||
|
||||
import warnings
|
||||
from io import BytesIO
|
||||
from math import ceil, log
|
||||
from typing import IO, NamedTuple
|
||||
|
||||
from . import BmpImagePlugin, Image, ImageFile, PngImagePlugin
|
||||
from ._binary import i16le as i16
|
||||
from ._binary import i32le as i32
|
||||
from ._binary import o8
|
||||
from ._binary import o16le as o16
|
||||
from ._binary import o32le as o32
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
_MAGIC = b"\0\0\1\0"
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
fp.write(_MAGIC) # (2+2)
|
||||
bmp = im.encoderinfo.get("bitmap_format") == "bmp"
|
||||
sizes = im.encoderinfo.get(
|
||||
"sizes",
|
||||
[(16, 16), (24, 24), (32, 32), (48, 48), (64, 64), (128, 128), (256, 256)],
|
||||
)
|
||||
frames = []
|
||||
provided_ims = [im] + im.encoderinfo.get("append_images", [])
|
||||
width, height = im.size
|
||||
for size in sorted(set(sizes)):
|
||||
if size[0] > width or size[1] > height or size[0] > 256 or size[1] > 256:
|
||||
continue
|
||||
|
||||
for provided_im in provided_ims:
|
||||
if provided_im.size != size:
|
||||
continue
|
||||
frames.append(provided_im)
|
||||
if bmp:
|
||||
bits = BmpImagePlugin.SAVE[provided_im.mode][1]
|
||||
bits_used = [bits]
|
||||
for other_im in provided_ims:
|
||||
if other_im.size != size:
|
||||
continue
|
||||
bits = BmpImagePlugin.SAVE[other_im.mode][1]
|
||||
if bits not in bits_used:
|
||||
# Another image has been supplied for this size
|
||||
# with a different bit depth
|
||||
frames.append(other_im)
|
||||
bits_used.append(bits)
|
||||
break
|
||||
else:
|
||||
# TODO: invent a more convenient method for proportional scalings
|
||||
frame = provided_im.copy()
|
||||
frame.thumbnail(size, Image.Resampling.LANCZOS, reducing_gap=None)
|
||||
frames.append(frame)
|
||||
fp.write(o16(len(frames))) # idCount(2)
|
||||
offset = fp.tell() + len(frames) * 16
|
||||
for frame in frames:
|
||||
width, height = frame.size
|
||||
# 0 means 256
|
||||
fp.write(o8(width if width < 256 else 0)) # bWidth(1)
|
||||
fp.write(o8(height if height < 256 else 0)) # bHeight(1)
|
||||
|
||||
bits, colors = BmpImagePlugin.SAVE[frame.mode][1:] if bmp else (32, 0)
|
||||
fp.write(o8(colors)) # bColorCount(1)
|
||||
fp.write(b"\0") # bReserved(1)
|
||||
fp.write(b"\0\0") # wPlanes(2)
|
||||
fp.write(o16(bits)) # wBitCount(2)
|
||||
|
||||
image_io = BytesIO()
|
||||
if bmp:
|
||||
frame.save(image_io, "dib")
|
||||
|
||||
if bits != 32:
|
||||
and_mask = Image.new("1", size)
|
||||
ImageFile._save(
|
||||
and_mask,
|
||||
image_io,
|
||||
[ImageFile._Tile("raw", (0, 0) + size, 0, ("1", 0, -1))],
|
||||
)
|
||||
else:
|
||||
frame.save(image_io, "png")
|
||||
image_io.seek(0)
|
||||
image_bytes = image_io.read()
|
||||
if bmp:
|
||||
image_bytes = image_bytes[:8] + o32(height * 2) + image_bytes[12:]
|
||||
bytes_len = len(image_bytes)
|
||||
fp.write(o32(bytes_len)) # dwBytesInRes(4)
|
||||
fp.write(o32(offset)) # dwImageOffset(4)
|
||||
current = fp.tell()
|
||||
fp.seek(offset)
|
||||
fp.write(image_bytes)
|
||||
offset = offset + bytes_len
|
||||
fp.seek(current)
|
||||
|
||||
|
||||
def _accept(prefix: bytes) -> bool:
|
||||
return prefix[:4] == _MAGIC
|
||||
|
||||
|
||||
class IconHeader(NamedTuple):
|
||||
width: int
|
||||
height: int
|
||||
nb_color: int
|
||||
reserved: int
|
||||
planes: int
|
||||
bpp: int
|
||||
size: int
|
||||
offset: int
|
||||
dim: tuple[int, int]
|
||||
square: int
|
||||
color_depth: int
|
||||
|
||||
|
||||
class IcoFile:
|
||||
def __init__(self, buf: IO[bytes]) -> None:
|
||||
"""
|
||||
Parse image from file-like object containing ico file data
|
||||
"""
|
||||
|
||||
# check magic
|
||||
s = buf.read(6)
|
||||
if not _accept(s):
|
||||
msg = "not an ICO file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
self.buf = buf
|
||||
self.entry = []
|
||||
|
||||
# Number of items in file
|
||||
self.nb_items = i16(s, 4)
|
||||
|
||||
# Get headers for each item
|
||||
for i in range(self.nb_items):
|
||||
s = buf.read(16)
|
||||
|
||||
# See Wikipedia
|
||||
width = s[0] or 256
|
||||
height = s[1] or 256
|
||||
|
||||
# No. of colors in image (0 if >=8bpp)
|
||||
nb_color = s[2]
|
||||
bpp = i16(s, 6)
|
||||
icon_header = IconHeader(
|
||||
width=width,
|
||||
height=height,
|
||||
nb_color=nb_color,
|
||||
reserved=s[3],
|
||||
planes=i16(s, 4),
|
||||
bpp=i16(s, 6),
|
||||
size=i32(s, 8),
|
||||
offset=i32(s, 12),
|
||||
dim=(width, height),
|
||||
square=width * height,
|
||||
# See Wikipedia notes about color depth.
|
||||
# We need this just to differ images with equal sizes
|
||||
color_depth=bpp or (nb_color != 0 and ceil(log(nb_color, 2))) or 256,
|
||||
)
|
||||
|
||||
self.entry.append(icon_header)
|
||||
|
||||
self.entry = sorted(self.entry, key=lambda x: x.color_depth)
|
||||
# ICO images are usually squares
|
||||
self.entry = sorted(self.entry, key=lambda x: x.square, reverse=True)
|
||||
|
||||
def sizes(self) -> set[tuple[int, int]]:
|
||||
"""
|
||||
Get a set of all available icon sizes and color depths.
|
||||
"""
|
||||
return {(h.width, h.height) for h in self.entry}
|
||||
|
||||
def getentryindex(self, size: tuple[int, int], bpp: int | bool = False) -> int:
|
||||
for i, h in enumerate(self.entry):
|
||||
if size == h.dim and (bpp is False or bpp == h.color_depth):
|
||||
return i
|
||||
return 0
|
||||
|
||||
def getimage(self, size: tuple[int, int], bpp: int | bool = False) -> Image.Image:
|
||||
"""
|
||||
Get an image from the icon
|
||||
"""
|
||||
return self.frame(self.getentryindex(size, bpp))
|
||||
|
||||
def frame(self, idx: int) -> Image.Image:
|
||||
"""
|
||||
Get an image from frame idx
|
||||
"""
|
||||
|
||||
header = self.entry[idx]
|
||||
|
||||
self.buf.seek(header.offset)
|
||||
data = self.buf.read(8)
|
||||
self.buf.seek(header.offset)
|
||||
|
||||
im: Image.Image
|
||||
if data[:8] == PngImagePlugin._MAGIC:
|
||||
# png frame
|
||||
im = PngImagePlugin.PngImageFile(self.buf)
|
||||
Image._decompression_bomb_check(im.size)
|
||||
else:
|
||||
# XOR + AND mask bmp frame
|
||||
im = BmpImagePlugin.DibImageFile(self.buf)
|
||||
Image._decompression_bomb_check(im.size)
|
||||
|
||||
# change tile dimension to only encompass XOR image
|
||||
im._size = (im.size[0], int(im.size[1] / 2))
|
||||
d, e, o, a = im.tile[0]
|
||||
im.tile[0] = ImageFile._Tile(d, (0, 0) + im.size, o, a)
|
||||
|
||||
# figure out where AND mask image starts
|
||||
if header.bpp == 32:
|
||||
# 32-bit color depth icon image allows semitransparent areas
|
||||
# PIL's DIB format ignores transparency bits, recover them.
|
||||
# The DIB is packed in BGRX byte order where X is the alpha
|
||||
# channel.
|
||||
|
||||
# Back up to start of bmp data
|
||||
self.buf.seek(o)
|
||||
# extract every 4th byte (eg. 3,7,11,15,...)
|
||||
alpha_bytes = self.buf.read(im.size[0] * im.size[1] * 4)[3::4]
|
||||
|
||||
# convert to an 8bpp grayscale image
|
||||
try:
|
||||
mask = Image.frombuffer(
|
||||
"L", # 8bpp
|
||||
im.size, # (w, h)
|
||||
alpha_bytes, # source chars
|
||||
"raw", # raw decoder
|
||||
("L", 0, -1), # 8bpp inverted, unpadded, reversed
|
||||
)
|
||||
except ValueError:
|
||||
if ImageFile.LOAD_TRUNCATED_IMAGES:
|
||||
mask = None
|
||||
else:
|
||||
raise
|
||||
else:
|
||||
# get AND image from end of bitmap
|
||||
w = im.size[0]
|
||||
if (w % 32) > 0:
|
||||
# bitmap row data is aligned to word boundaries
|
||||
w += 32 - (im.size[0] % 32)
|
||||
|
||||
# the total mask data is
|
||||
# padded row size * height / bits per char
|
||||
|
||||
total_bytes = int((w * im.size[1]) / 8)
|
||||
and_mask_offset = header.offset + header.size - total_bytes
|
||||
|
||||
self.buf.seek(and_mask_offset)
|
||||
mask_data = self.buf.read(total_bytes)
|
||||
|
||||
# convert raw data to image
|
||||
try:
|
||||
mask = Image.frombuffer(
|
||||
"1", # 1 bpp
|
||||
im.size, # (w, h)
|
||||
mask_data, # source chars
|
||||
"raw", # raw decoder
|
||||
("1;I", int(w / 8), -1), # 1bpp inverted, padded, reversed
|
||||
)
|
||||
except ValueError:
|
||||
if ImageFile.LOAD_TRUNCATED_IMAGES:
|
||||
mask = None
|
||||
else:
|
||||
raise
|
||||
|
||||
# now we have two images, im is XOR image and mask is AND image
|
||||
|
||||
# apply mask image as alpha channel
|
||||
if mask:
|
||||
im = im.convert("RGBA")
|
||||
im.putalpha(mask)
|
||||
|
||||
return im
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for Windows Icon files.
|
||||
|
||||
|
||||
class IcoImageFile(ImageFile.ImageFile):
|
||||
"""
|
||||
PIL read-only image support for Microsoft Windows .ico files.
|
||||
|
||||
By default the largest resolution image in the file will be loaded. This
|
||||
can be changed by altering the 'size' attribute before calling 'load'.
|
||||
|
||||
The info dictionary has a key 'sizes' that is a list of the sizes available
|
||||
in the icon file.
|
||||
|
||||
Handles classic, XP and Vista icon formats.
|
||||
|
||||
When saving, PNG compression is used. Support for this was only added in
|
||||
Windows Vista. If you are unable to view the icon in Windows, convert the
|
||||
image to "RGBA" mode before saving.
|
||||
|
||||
This plugin is a refactored version of Win32IconImagePlugin by Bryan Davis
|
||||
<casadebender@gmail.com>.
|
||||
https://code.google.com/archive/p/casadebender/wikis/Win32IconImagePlugin.wiki
|
||||
"""
|
||||
|
||||
format = "ICO"
|
||||
format_description = "Windows Icon"
|
||||
|
||||
def _open(self) -> None:
|
||||
self.ico = IcoFile(self.fp)
|
||||
self.info["sizes"] = self.ico.sizes()
|
||||
self.size = self.ico.entry[0].dim
|
||||
self.load()
|
||||
|
||||
@property
|
||||
def size(self) -> tuple[int, int]:
|
||||
return self._size
|
||||
|
||||
@size.setter
|
||||
def size(self, value: tuple[int, int]) -> None:
|
||||
if value not in self.info["sizes"]:
|
||||
msg = "This is not one of the allowed sizes of this image"
|
||||
raise ValueError(msg)
|
||||
self._size = value
|
||||
|
||||
def load(self) -> Image.core.PixelAccess | None:
|
||||
if self._im is not None and self.im.size == self.size:
|
||||
# Already loaded
|
||||
return Image.Image.load(self)
|
||||
im = self.ico.getimage(self.size)
|
||||
# if tile is PNG, it won't really be loaded yet
|
||||
im.load()
|
||||
self.im = im.im
|
||||
self._mode = im.mode
|
||||
if im.palette:
|
||||
self.palette = im.palette
|
||||
if im.size != self.size:
|
||||
warnings.warn("Image was not the expected size")
|
||||
|
||||
index = self.ico.getentryindex(self.size)
|
||||
sizes = list(self.info["sizes"])
|
||||
sizes[index] = im.size
|
||||
self.info["sizes"] = set(sizes)
|
||||
|
||||
self.size = im.size
|
||||
return None
|
||||
|
||||
def load_seek(self, pos: int) -> None:
|
||||
# Flag the ImageFile.Parser so that it
|
||||
# just does all the decode at the end.
|
||||
pass
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
|
||||
|
||||
Image.register_open(IcoImageFile.format, IcoImageFile, _accept)
|
||||
Image.register_save(IcoImageFile.format, _save)
|
||||
Image.register_extension(IcoImageFile.format, ".ico")
|
||||
|
||||
Image.register_mime(IcoImageFile.format, "image/x-icon")
|
||||
386
testclass/lib/python3.12/site-packages/PIL/ImImagePlugin.py
Normal file
386
testclass/lib/python3.12/site-packages/PIL/ImImagePlugin.py
Normal file
@ -0,0 +1,386 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# IFUNC IM file handling for PIL
|
||||
#
|
||||
# history:
|
||||
# 1995-09-01 fl Created.
|
||||
# 1997-01-03 fl Save palette images
|
||||
# 1997-01-08 fl Added sequence support
|
||||
# 1997-01-23 fl Added P and RGB save support
|
||||
# 1997-05-31 fl Read floating point images
|
||||
# 1997-06-22 fl Save floating point images
|
||||
# 1997-08-27 fl Read and save 1-bit images
|
||||
# 1998-06-25 fl Added support for RGB+LUT images
|
||||
# 1998-07-02 fl Added support for YCC images
|
||||
# 1998-07-15 fl Renamed offset attribute to avoid name clash
|
||||
# 1998-12-29 fl Added I;16 support
|
||||
# 2001-02-17 fl Use 're' instead of 'regex' (Python 2.1) (0.7)
|
||||
# 2003-09-26 fl Added LA/PA support
|
||||
#
|
||||
# Copyright (c) 1997-2003 by Secret Labs AB.
|
||||
# Copyright (c) 1995-2001 by Fredrik Lundh.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import os
|
||||
import re
|
||||
from typing import IO, Any
|
||||
|
||||
from . import Image, ImageFile, ImagePalette
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Standard tags
|
||||
|
||||
COMMENT = "Comment"
|
||||
DATE = "Date"
|
||||
EQUIPMENT = "Digitalization equipment"
|
||||
FRAMES = "File size (no of images)"
|
||||
LUT = "Lut"
|
||||
NAME = "Name"
|
||||
SCALE = "Scale (x,y)"
|
||||
SIZE = "Image size (x*y)"
|
||||
MODE = "Image type"
|
||||
|
||||
TAGS = {
|
||||
COMMENT: 0,
|
||||
DATE: 0,
|
||||
EQUIPMENT: 0,
|
||||
FRAMES: 0,
|
||||
LUT: 0,
|
||||
NAME: 0,
|
||||
SCALE: 0,
|
||||
SIZE: 0,
|
||||
MODE: 0,
|
||||
}
|
||||
|
||||
OPEN = {
|
||||
# ifunc93/p3cfunc formats
|
||||
"0 1 image": ("1", "1"),
|
||||
"L 1 image": ("1", "1"),
|
||||
"Greyscale image": ("L", "L"),
|
||||
"Grayscale image": ("L", "L"),
|
||||
"RGB image": ("RGB", "RGB;L"),
|
||||
"RLB image": ("RGB", "RLB"),
|
||||
"RYB image": ("RGB", "RLB"),
|
||||
"B1 image": ("1", "1"),
|
||||
"B2 image": ("P", "P;2"),
|
||||
"B4 image": ("P", "P;4"),
|
||||
"X 24 image": ("RGB", "RGB"),
|
||||
"L 32 S image": ("I", "I;32"),
|
||||
"L 32 F image": ("F", "F;32"),
|
||||
# old p3cfunc formats
|
||||
"RGB3 image": ("RGB", "RGB;T"),
|
||||
"RYB3 image": ("RGB", "RYB;T"),
|
||||
# extensions
|
||||
"LA image": ("LA", "LA;L"),
|
||||
"PA image": ("LA", "PA;L"),
|
||||
"RGBA image": ("RGBA", "RGBA;L"),
|
||||
"RGBX image": ("RGB", "RGBX;L"),
|
||||
"CMYK image": ("CMYK", "CMYK;L"),
|
||||
"YCC image": ("YCbCr", "YCbCr;L"),
|
||||
}
|
||||
|
||||
# ifunc95 extensions
|
||||
for i in ["8", "8S", "16", "16S", "32", "32F"]:
|
||||
OPEN[f"L {i} image"] = ("F", f"F;{i}")
|
||||
OPEN[f"L*{i} image"] = ("F", f"F;{i}")
|
||||
for i in ["16", "16L", "16B"]:
|
||||
OPEN[f"L {i} image"] = (f"I;{i}", f"I;{i}")
|
||||
OPEN[f"L*{i} image"] = (f"I;{i}", f"I;{i}")
|
||||
for i in ["32S"]:
|
||||
OPEN[f"L {i} image"] = ("I", f"I;{i}")
|
||||
OPEN[f"L*{i} image"] = ("I", f"I;{i}")
|
||||
for j in range(2, 33):
|
||||
OPEN[f"L*{j} image"] = ("F", f"F;{j}")
|
||||
|
||||
|
||||
# --------------------------------------------------------------------
|
||||
# Read IM directory
|
||||
|
||||
split = re.compile(rb"^([A-Za-z][^:]*):[ \t]*(.*)[ \t]*$")
|
||||
|
||||
|
||||
def number(s: Any) -> float:
|
||||
try:
|
||||
return int(s)
|
||||
except ValueError:
|
||||
return float(s)
|
||||
|
||||
|
||||
##
|
||||
# Image plugin for the IFUNC IM file format.
|
||||
|
||||
|
||||
class ImImageFile(ImageFile.ImageFile):
|
||||
format = "IM"
|
||||
format_description = "IFUNC Image Memory"
|
||||
_close_exclusive_fp_after_loading = False
|
||||
|
||||
def _open(self) -> None:
|
||||
# Quick rejection: if there's not an LF among the first
|
||||
# 100 bytes, this is (probably) not a text header.
|
||||
|
||||
if b"\n" not in self.fp.read(100):
|
||||
msg = "not an IM file"
|
||||
raise SyntaxError(msg)
|
||||
self.fp.seek(0)
|
||||
|
||||
n = 0
|
||||
|
||||
# Default values
|
||||
self.info[MODE] = "L"
|
||||
self.info[SIZE] = (512, 512)
|
||||
self.info[FRAMES] = 1
|
||||
|
||||
self.rawmode = "L"
|
||||
|
||||
while True:
|
||||
s = self.fp.read(1)
|
||||
|
||||
# Some versions of IFUNC uses \n\r instead of \r\n...
|
||||
if s == b"\r":
|
||||
continue
|
||||
|
||||
if not s or s == b"\0" or s == b"\x1A":
|
||||
break
|
||||
|
||||
# FIXME: this may read whole file if not a text file
|
||||
s = s + self.fp.readline()
|
||||
|
||||
if len(s) > 100:
|
||||
msg = "not an IM file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
if s[-2:] == b"\r\n":
|
||||
s = s[:-2]
|
||||
elif s[-1:] == b"\n":
|
||||
s = s[:-1]
|
||||
|
||||
try:
|
||||
m = split.match(s)
|
||||
except re.error as e:
|
||||
msg = "not an IM file"
|
||||
raise SyntaxError(msg) from e
|
||||
|
||||
if m:
|
||||
k, v = m.group(1, 2)
|
||||
|
||||
# Don't know if this is the correct encoding,
|
||||
# but a decent guess (I guess)
|
||||
k = k.decode("latin-1", "replace")
|
||||
v = v.decode("latin-1", "replace")
|
||||
|
||||
# Convert value as appropriate
|
||||
if k in [FRAMES, SCALE, SIZE]:
|
||||
v = v.replace("*", ",")
|
||||
v = tuple(map(number, v.split(",")))
|
||||
if len(v) == 1:
|
||||
v = v[0]
|
||||
elif k == MODE and v in OPEN:
|
||||
v, self.rawmode = OPEN[v]
|
||||
|
||||
# Add to dictionary. Note that COMMENT tags are
|
||||
# combined into a list of strings.
|
||||
if k == COMMENT:
|
||||
if k in self.info:
|
||||
self.info[k].append(v)
|
||||
else:
|
||||
self.info[k] = [v]
|
||||
else:
|
||||
self.info[k] = v
|
||||
|
||||
if k in TAGS:
|
||||
n += 1
|
||||
|
||||
else:
|
||||
msg = f"Syntax error in IM header: {s.decode('ascii', 'replace')}"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
if not n:
|
||||
msg = "Not an IM file"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
# Basic attributes
|
||||
self._size = self.info[SIZE]
|
||||
self._mode = self.info[MODE]
|
||||
|
||||
# Skip forward to start of image data
|
||||
while s and s[:1] != b"\x1A":
|
||||
s = self.fp.read(1)
|
||||
if not s:
|
||||
msg = "File truncated"
|
||||
raise SyntaxError(msg)
|
||||
|
||||
if LUT in self.info:
|
||||
# convert lookup table to palette or lut attribute
|
||||
palette = self.fp.read(768)
|
||||
greyscale = 1 # greyscale palette
|
||||
linear = 1 # linear greyscale palette
|
||||
for i in range(256):
|
||||
if palette[i] == palette[i + 256] == palette[i + 512]:
|
||||
if palette[i] != i:
|
||||
linear = 0
|
||||
else:
|
||||
greyscale = 0
|
||||
if self.mode in ["L", "LA", "P", "PA"]:
|
||||
if greyscale:
|
||||
if not linear:
|
||||
self.lut = list(palette[:256])
|
||||
else:
|
||||
if self.mode in ["L", "P"]:
|
||||
self._mode = self.rawmode = "P"
|
||||
elif self.mode in ["LA", "PA"]:
|
||||
self._mode = "PA"
|
||||
self.rawmode = "PA;L"
|
||||
self.palette = ImagePalette.raw("RGB;L", palette)
|
||||
elif self.mode == "RGB":
|
||||
if not greyscale or not linear:
|
||||
self.lut = list(palette)
|
||||
|
||||
self.frame = 0
|
||||
|
||||
self.__offset = offs = self.fp.tell()
|
||||
|
||||
self._fp = self.fp # FIXME: hack
|
||||
|
||||
if self.rawmode[:2] == "F;":
|
||||
# ifunc95 formats
|
||||
try:
|
||||
# use bit decoder (if necessary)
|
||||
bits = int(self.rawmode[2:])
|
||||
if bits not in [8, 16, 32]:
|
||||
self.tile = [
|
||||
ImageFile._Tile(
|
||||
"bit", (0, 0) + self.size, offs, (bits, 8, 3, 0, -1)
|
||||
)
|
||||
]
|
||||
return
|
||||
except ValueError:
|
||||
pass
|
||||
|
||||
if self.rawmode in ["RGB;T", "RYB;T"]:
|
||||
# Old LabEye/3PC files. Would be very surprised if anyone
|
||||
# ever stumbled upon such a file ;-)
|
||||
size = self.size[0] * self.size[1]
|
||||
self.tile = [
|
||||
ImageFile._Tile("raw", (0, 0) + self.size, offs, ("G", 0, -1)),
|
||||
ImageFile._Tile("raw", (0, 0) + self.size, offs + size, ("R", 0, -1)),
|
||||
ImageFile._Tile(
|
||||
"raw", (0, 0) + self.size, offs + 2 * size, ("B", 0, -1)
|
||||
),
|
||||
]
|
||||
else:
|
||||
# LabEye/IFUNC files
|
||||
self.tile = [
|
||||
ImageFile._Tile("raw", (0, 0) + self.size, offs, (self.rawmode, 0, -1))
|
||||
]
|
||||
|
||||
@property
|
||||
def n_frames(self) -> int:
|
||||
return self.info[FRAMES]
|
||||
|
||||
@property
|
||||
def is_animated(self) -> bool:
|
||||
return self.info[FRAMES] > 1
|
||||
|
||||
def seek(self, frame: int) -> None:
|
||||
if not self._seek_check(frame):
|
||||
return
|
||||
|
||||
self.frame = frame
|
||||
|
||||
if self.mode == "1":
|
||||
bits = 1
|
||||
else:
|
||||
bits = 8 * len(self.mode)
|
||||
|
||||
size = ((self.size[0] * bits + 7) // 8) * self.size[1]
|
||||
offs = self.__offset + frame * size
|
||||
|
||||
self.fp = self._fp
|
||||
|
||||
self.tile = [
|
||||
ImageFile._Tile("raw", (0, 0) + self.size, offs, (self.rawmode, 0, -1))
|
||||
]
|
||||
|
||||
def tell(self) -> int:
|
||||
return self.frame
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
# Save IM files
|
||||
|
||||
|
||||
SAVE = {
|
||||
# mode: (im type, raw mode)
|
||||
"1": ("0 1", "1"),
|
||||
"L": ("Greyscale", "L"),
|
||||
"LA": ("LA", "LA;L"),
|
||||
"P": ("Greyscale", "P"),
|
||||
"PA": ("LA", "PA;L"),
|
||||
"I": ("L 32S", "I;32S"),
|
||||
"I;16": ("L 16", "I;16"),
|
||||
"I;16L": ("L 16L", "I;16L"),
|
||||
"I;16B": ("L 16B", "I;16B"),
|
||||
"F": ("L 32F", "F;32F"),
|
||||
"RGB": ("RGB", "RGB;L"),
|
||||
"RGBA": ("RGBA", "RGBA;L"),
|
||||
"RGBX": ("RGBX", "RGBX;L"),
|
||||
"CMYK": ("CMYK", "CMYK;L"),
|
||||
"YCbCr": ("YCC", "YCbCr;L"),
|
||||
}
|
||||
|
||||
|
||||
def _save(im: Image.Image, fp: IO[bytes], filename: str | bytes) -> None:
|
||||
try:
|
||||
image_type, rawmode = SAVE[im.mode]
|
||||
except KeyError as e:
|
||||
msg = f"Cannot save {im.mode} images as IM"
|
||||
raise ValueError(msg) from e
|
||||
|
||||
frames = im.encoderinfo.get("frames", 1)
|
||||
|
||||
fp.write(f"Image type: {image_type} image\r\n".encode("ascii"))
|
||||
if filename:
|
||||
# Each line must be 100 characters or less,
|
||||
# or: SyntaxError("not an IM file")
|
||||
# 8 characters are used for "Name: " and "\r\n"
|
||||
# Keep just the filename, ditch the potentially overlong path
|
||||
if isinstance(filename, bytes):
|
||||
filename = filename.decode("ascii")
|
||||
name, ext = os.path.splitext(os.path.basename(filename))
|
||||
name = "".join([name[: 92 - len(ext)], ext])
|
||||
|
||||
fp.write(f"Name: {name}\r\n".encode("ascii"))
|
||||
fp.write(f"Image size (x*y): {im.size[0]}*{im.size[1]}\r\n".encode("ascii"))
|
||||
fp.write(f"File size (no of images): {frames}\r\n".encode("ascii"))
|
||||
if im.mode in ["P", "PA"]:
|
||||
fp.write(b"Lut: 1\r\n")
|
||||
fp.write(b"\000" * (511 - fp.tell()) + b"\032")
|
||||
if im.mode in ["P", "PA"]:
|
||||
im_palette = im.im.getpalette("RGB", "RGB;L")
|
||||
colors = len(im_palette) // 3
|
||||
palette = b""
|
||||
for i in range(3):
|
||||
palette += im_palette[colors * i : colors * (i + 1)]
|
||||
palette += b"\x00" * (256 - colors)
|
||||
fp.write(palette) # 768 bytes
|
||||
ImageFile._save(
|
||||
im, fp, [ImageFile._Tile("raw", (0, 0) + im.size, 0, (rawmode, 0, -1))]
|
||||
)
|
||||
|
||||
|
||||
#
|
||||
# --------------------------------------------------------------------
|
||||
# Registry
|
||||
|
||||
|
||||
Image.register_open(ImImageFile.format, ImImageFile)
|
||||
Image.register_save(ImImageFile.format, _save)
|
||||
|
||||
Image.register_extension(ImImageFile.format, ".im")
|
||||
4197
testclass/lib/python3.12/site-packages/PIL/Image.py
Normal file
4197
testclass/lib/python3.12/site-packages/PIL/Image.py
Normal file
File diff suppressed because it is too large
Load Diff
311
testclass/lib/python3.12/site-packages/PIL/ImageChops.py
Normal file
311
testclass/lib/python3.12/site-packages/PIL/ImageChops.py
Normal file
@ -0,0 +1,311 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# standard channel operations
|
||||
#
|
||||
# History:
|
||||
# 1996-03-24 fl Created
|
||||
# 1996-08-13 fl Added logical operations (for "1" images)
|
||||
# 2000-10-12 fl Added offset method (from Image.py)
|
||||
#
|
||||
# Copyright (c) 1997-2000 by Secret Labs AB
|
||||
# Copyright (c) 1996-2000 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
from . import Image
|
||||
|
||||
|
||||
def constant(image: Image.Image, value: int) -> Image.Image:
|
||||
"""Fill a channel with a given gray level.
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
return Image.new("L", image.size, value)
|
||||
|
||||
|
||||
def duplicate(image: Image.Image) -> Image.Image:
|
||||
"""Copy a channel. Alias for :py:meth:`PIL.Image.Image.copy`.
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
return image.copy()
|
||||
|
||||
|
||||
def invert(image: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Invert an image (channel). ::
|
||||
|
||||
out = MAX - image
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image.load()
|
||||
return image._new(image.im.chop_invert())
|
||||
|
||||
|
||||
def lighter(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Compares the two images, pixel by pixel, and returns a new image containing
|
||||
the lighter values. ::
|
||||
|
||||
out = max(image1, image2)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_lighter(image2.im))
|
||||
|
||||
|
||||
def darker(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Compares the two images, pixel by pixel, and returns a new image containing
|
||||
the darker values. ::
|
||||
|
||||
out = min(image1, image2)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_darker(image2.im))
|
||||
|
||||
|
||||
def difference(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Returns the absolute value of the pixel-by-pixel difference between the two
|
||||
images. ::
|
||||
|
||||
out = abs(image1 - image2)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_difference(image2.im))
|
||||
|
||||
|
||||
def multiply(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Superimposes two images on top of each other.
|
||||
|
||||
If you multiply an image with a solid black image, the result is black. If
|
||||
you multiply with a solid white image, the image is unaffected. ::
|
||||
|
||||
out = image1 * image2 / MAX
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_multiply(image2.im))
|
||||
|
||||
|
||||
def screen(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Superimposes two inverted images on top of each other. ::
|
||||
|
||||
out = MAX - ((MAX - image1) * (MAX - image2) / MAX)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_screen(image2.im))
|
||||
|
||||
|
||||
def soft_light(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Superimposes two images on top of each other using the Soft Light algorithm
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_soft_light(image2.im))
|
||||
|
||||
|
||||
def hard_light(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Superimposes two images on top of each other using the Hard Light algorithm
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_hard_light(image2.im))
|
||||
|
||||
|
||||
def overlay(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""
|
||||
Superimposes two images on top of each other using the Overlay algorithm
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_overlay(image2.im))
|
||||
|
||||
|
||||
def add(
|
||||
image1: Image.Image, image2: Image.Image, scale: float = 1.0, offset: float = 0
|
||||
) -> Image.Image:
|
||||
"""
|
||||
Adds two images, dividing the result by scale and adding the
|
||||
offset. If omitted, scale defaults to 1.0, and offset to 0.0. ::
|
||||
|
||||
out = ((image1 + image2) / scale + offset)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_add(image2.im, scale, offset))
|
||||
|
||||
|
||||
def subtract(
|
||||
image1: Image.Image, image2: Image.Image, scale: float = 1.0, offset: float = 0
|
||||
) -> Image.Image:
|
||||
"""
|
||||
Subtracts two images, dividing the result by scale and adding the offset.
|
||||
If omitted, scale defaults to 1.0, and offset to 0.0. ::
|
||||
|
||||
out = ((image1 - image2) / scale + offset)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_subtract(image2.im, scale, offset))
|
||||
|
||||
|
||||
def add_modulo(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""Add two images, without clipping the result. ::
|
||||
|
||||
out = ((image1 + image2) % MAX)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_add_modulo(image2.im))
|
||||
|
||||
|
||||
def subtract_modulo(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""Subtract two images, without clipping the result. ::
|
||||
|
||||
out = ((image1 - image2) % MAX)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_subtract_modulo(image2.im))
|
||||
|
||||
|
||||
def logical_and(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""Logical AND between two images.
|
||||
|
||||
Both of the images must have mode "1". If you would like to perform a
|
||||
logical AND on an image with a mode other than "1", try
|
||||
:py:meth:`~PIL.ImageChops.multiply` instead, using a black-and-white mask
|
||||
as the second image. ::
|
||||
|
||||
out = ((image1 and image2) % MAX)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_and(image2.im))
|
||||
|
||||
|
||||
def logical_or(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""Logical OR between two images.
|
||||
|
||||
Both of the images must have mode "1". ::
|
||||
|
||||
out = ((image1 or image2) % MAX)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_or(image2.im))
|
||||
|
||||
|
||||
def logical_xor(image1: Image.Image, image2: Image.Image) -> Image.Image:
|
||||
"""Logical XOR between two images.
|
||||
|
||||
Both of the images must have mode "1". ::
|
||||
|
||||
out = ((bool(image1) != bool(image2)) % MAX)
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
image1.load()
|
||||
image2.load()
|
||||
return image1._new(image1.im.chop_xor(image2.im))
|
||||
|
||||
|
||||
def blend(image1: Image.Image, image2: Image.Image, alpha: float) -> Image.Image:
|
||||
"""Blend images using constant transparency weight. Alias for
|
||||
:py:func:`PIL.Image.blend`.
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
return Image.blend(image1, image2, alpha)
|
||||
|
||||
|
||||
def composite(
|
||||
image1: Image.Image, image2: Image.Image, mask: Image.Image
|
||||
) -> Image.Image:
|
||||
"""Create composite using transparency mask. Alias for
|
||||
:py:func:`PIL.Image.composite`.
|
||||
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
return Image.composite(image1, image2, mask)
|
||||
|
||||
|
||||
def offset(image: Image.Image, xoffset: int, yoffset: int | None = None) -> Image.Image:
|
||||
"""Returns a copy of the image where data has been offset by the given
|
||||
distances. Data wraps around the edges. If ``yoffset`` is omitted, it
|
||||
is assumed to be equal to ``xoffset``.
|
||||
|
||||
:param image: Input image.
|
||||
:param xoffset: The horizontal distance.
|
||||
:param yoffset: The vertical distance. If omitted, both
|
||||
distances are set to the same value.
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
|
||||
if yoffset is None:
|
||||
yoffset = xoffset
|
||||
image.load()
|
||||
return image._new(image.im.offset(xoffset, yoffset))
|
||||
1125
testclass/lib/python3.12/site-packages/PIL/ImageCms.py
Normal file
1125
testclass/lib/python3.12/site-packages/PIL/ImageCms.py
Normal file
File diff suppressed because it is too large
Load Diff
320
testclass/lib/python3.12/site-packages/PIL/ImageColor.py
Normal file
320
testclass/lib/python3.12/site-packages/PIL/ImageColor.py
Normal file
@ -0,0 +1,320 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# map CSS3-style colour description strings to RGB
|
||||
#
|
||||
# History:
|
||||
# 2002-10-24 fl Added support for CSS-style color strings
|
||||
# 2002-12-15 fl Added RGBA support
|
||||
# 2004-03-27 fl Fixed remaining int() problems for Python 1.5.2
|
||||
# 2004-07-19 fl Fixed gray/grey spelling issues
|
||||
# 2009-03-05 fl Fixed rounding error in grayscale calculation
|
||||
#
|
||||
# Copyright (c) 2002-2004 by Secret Labs AB
|
||||
# Copyright (c) 2002-2004 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
import re
|
||||
from functools import lru_cache
|
||||
|
||||
from . import Image
|
||||
|
||||
|
||||
@lru_cache
|
||||
def getrgb(color: str) -> tuple[int, int, int] | tuple[int, int, int, int]:
|
||||
"""
|
||||
Convert a color string to an RGB or RGBA tuple. If the string cannot be
|
||||
parsed, this function raises a :py:exc:`ValueError` exception.
|
||||
|
||||
.. versionadded:: 1.1.4
|
||||
|
||||
:param color: A color string
|
||||
:return: ``(red, green, blue[, alpha])``
|
||||
"""
|
||||
if len(color) > 100:
|
||||
msg = "color specifier is too long"
|
||||
raise ValueError(msg)
|
||||
color = color.lower()
|
||||
|
||||
rgb = colormap.get(color, None)
|
||||
if rgb:
|
||||
if isinstance(rgb, tuple):
|
||||
return rgb
|
||||
rgb_tuple = getrgb(rgb)
|
||||
assert len(rgb_tuple) == 3
|
||||
colormap[color] = rgb_tuple
|
||||
return rgb_tuple
|
||||
|
||||
# check for known string formats
|
||||
if re.match("#[a-f0-9]{3}$", color):
|
||||
return int(color[1] * 2, 16), int(color[2] * 2, 16), int(color[3] * 2, 16)
|
||||
|
||||
if re.match("#[a-f0-9]{4}$", color):
|
||||
return (
|
||||
int(color[1] * 2, 16),
|
||||
int(color[2] * 2, 16),
|
||||
int(color[3] * 2, 16),
|
||||
int(color[4] * 2, 16),
|
||||
)
|
||||
|
||||
if re.match("#[a-f0-9]{6}$", color):
|
||||
return int(color[1:3], 16), int(color[3:5], 16), int(color[5:7], 16)
|
||||
|
||||
if re.match("#[a-f0-9]{8}$", color):
|
||||
return (
|
||||
int(color[1:3], 16),
|
||||
int(color[3:5], 16),
|
||||
int(color[5:7], 16),
|
||||
int(color[7:9], 16),
|
||||
)
|
||||
|
||||
m = re.match(r"rgb\(\s*(\d+)\s*,\s*(\d+)\s*,\s*(\d+)\s*\)$", color)
|
||||
if m:
|
||||
return int(m.group(1)), int(m.group(2)), int(m.group(3))
|
||||
|
||||
m = re.match(r"rgb\(\s*(\d+)%\s*,\s*(\d+)%\s*,\s*(\d+)%\s*\)$", color)
|
||||
if m:
|
||||
return (
|
||||
int((int(m.group(1)) * 255) / 100.0 + 0.5),
|
||||
int((int(m.group(2)) * 255) / 100.0 + 0.5),
|
||||
int((int(m.group(3)) * 255) / 100.0 + 0.5),
|
||||
)
|
||||
|
||||
m = re.match(
|
||||
r"hsl\(\s*(\d+\.?\d*)\s*,\s*(\d+\.?\d*)%\s*,\s*(\d+\.?\d*)%\s*\)$", color
|
||||
)
|
||||
if m:
|
||||
from colorsys import hls_to_rgb
|
||||
|
||||
rgb_floats = hls_to_rgb(
|
||||
float(m.group(1)) / 360.0,
|
||||
float(m.group(3)) / 100.0,
|
||||
float(m.group(2)) / 100.0,
|
||||
)
|
||||
return (
|
||||
int(rgb_floats[0] * 255 + 0.5),
|
||||
int(rgb_floats[1] * 255 + 0.5),
|
||||
int(rgb_floats[2] * 255 + 0.5),
|
||||
)
|
||||
|
||||
m = re.match(
|
||||
r"hs[bv]\(\s*(\d+\.?\d*)\s*,\s*(\d+\.?\d*)%\s*,\s*(\d+\.?\d*)%\s*\)$", color
|
||||
)
|
||||
if m:
|
||||
from colorsys import hsv_to_rgb
|
||||
|
||||
rgb_floats = hsv_to_rgb(
|
||||
float(m.group(1)) / 360.0,
|
||||
float(m.group(2)) / 100.0,
|
||||
float(m.group(3)) / 100.0,
|
||||
)
|
||||
return (
|
||||
int(rgb_floats[0] * 255 + 0.5),
|
||||
int(rgb_floats[1] * 255 + 0.5),
|
||||
int(rgb_floats[2] * 255 + 0.5),
|
||||
)
|
||||
|
||||
m = re.match(r"rgba\(\s*(\d+)\s*,\s*(\d+)\s*,\s*(\d+)\s*,\s*(\d+)\s*\)$", color)
|
||||
if m:
|
||||
return int(m.group(1)), int(m.group(2)), int(m.group(3)), int(m.group(4))
|
||||
msg = f"unknown color specifier: {repr(color)}"
|
||||
raise ValueError(msg)
|
||||
|
||||
|
||||
@lru_cache
|
||||
def getcolor(color: str, mode: str) -> int | tuple[int, ...]:
|
||||
"""
|
||||
Same as :py:func:`~PIL.ImageColor.getrgb` for most modes. However, if
|
||||
``mode`` is HSV, converts the RGB value to a HSV value, or if ``mode`` is
|
||||
not color or a palette image, converts the RGB value to a grayscale value.
|
||||
If the string cannot be parsed, this function raises a :py:exc:`ValueError`
|
||||
exception.
|
||||
|
||||
.. versionadded:: 1.1.4
|
||||
|
||||
:param color: A color string
|
||||
:param mode: Convert result to this mode
|
||||
:return: ``graylevel, (graylevel, alpha) or (red, green, blue[, alpha])``
|
||||
"""
|
||||
# same as getrgb, but converts the result to the given mode
|
||||
rgb, alpha = getrgb(color), 255
|
||||
if len(rgb) == 4:
|
||||
alpha = rgb[3]
|
||||
rgb = rgb[:3]
|
||||
|
||||
if mode == "HSV":
|
||||
from colorsys import rgb_to_hsv
|
||||
|
||||
r, g, b = rgb
|
||||
h, s, v = rgb_to_hsv(r / 255, g / 255, b / 255)
|
||||
return int(h * 255), int(s * 255), int(v * 255)
|
||||
elif Image.getmodebase(mode) == "L":
|
||||
r, g, b = rgb
|
||||
# ITU-R Recommendation 601-2 for nonlinear RGB
|
||||
# scaled to 24 bits to match the convert's implementation.
|
||||
graylevel = (r * 19595 + g * 38470 + b * 7471 + 0x8000) >> 16
|
||||
if mode[-1] == "A":
|
||||
return graylevel, alpha
|
||||
return graylevel
|
||||
elif mode[-1] == "A":
|
||||
return rgb + (alpha,)
|
||||
return rgb
|
||||
|
||||
|
||||
colormap: dict[str, str | tuple[int, int, int]] = {
|
||||
# X11 colour table from https://drafts.csswg.org/css-color-4/, with
|
||||
# gray/grey spelling issues fixed. This is a superset of HTML 4.0
|
||||
# colour names used in CSS 1.
|
||||
"aliceblue": "#f0f8ff",
|
||||
"antiquewhite": "#faebd7",
|
||||
"aqua": "#00ffff",
|
||||
"aquamarine": "#7fffd4",
|
||||
"azure": "#f0ffff",
|
||||
"beige": "#f5f5dc",
|
||||
"bisque": "#ffe4c4",
|
||||
"black": "#000000",
|
||||
"blanchedalmond": "#ffebcd",
|
||||
"blue": "#0000ff",
|
||||
"blueviolet": "#8a2be2",
|
||||
"brown": "#a52a2a",
|
||||
"burlywood": "#deb887",
|
||||
"cadetblue": "#5f9ea0",
|
||||
"chartreuse": "#7fff00",
|
||||
"chocolate": "#d2691e",
|
||||
"coral": "#ff7f50",
|
||||
"cornflowerblue": "#6495ed",
|
||||
"cornsilk": "#fff8dc",
|
||||
"crimson": "#dc143c",
|
||||
"cyan": "#00ffff",
|
||||
"darkblue": "#00008b",
|
||||
"darkcyan": "#008b8b",
|
||||
"darkgoldenrod": "#b8860b",
|
||||
"darkgray": "#a9a9a9",
|
||||
"darkgrey": "#a9a9a9",
|
||||
"darkgreen": "#006400",
|
||||
"darkkhaki": "#bdb76b",
|
||||
"darkmagenta": "#8b008b",
|
||||
"darkolivegreen": "#556b2f",
|
||||
"darkorange": "#ff8c00",
|
||||
"darkorchid": "#9932cc",
|
||||
"darkred": "#8b0000",
|
||||
"darksalmon": "#e9967a",
|
||||
"darkseagreen": "#8fbc8f",
|
||||
"darkslateblue": "#483d8b",
|
||||
"darkslategray": "#2f4f4f",
|
||||
"darkslategrey": "#2f4f4f",
|
||||
"darkturquoise": "#00ced1",
|
||||
"darkviolet": "#9400d3",
|
||||
"deeppink": "#ff1493",
|
||||
"deepskyblue": "#00bfff",
|
||||
"dimgray": "#696969",
|
||||
"dimgrey": "#696969",
|
||||
"dodgerblue": "#1e90ff",
|
||||
"firebrick": "#b22222",
|
||||
"floralwhite": "#fffaf0",
|
||||
"forestgreen": "#228b22",
|
||||
"fuchsia": "#ff00ff",
|
||||
"gainsboro": "#dcdcdc",
|
||||
"ghostwhite": "#f8f8ff",
|
||||
"gold": "#ffd700",
|
||||
"goldenrod": "#daa520",
|
||||
"gray": "#808080",
|
||||
"grey": "#808080",
|
||||
"green": "#008000",
|
||||
"greenyellow": "#adff2f",
|
||||
"honeydew": "#f0fff0",
|
||||
"hotpink": "#ff69b4",
|
||||
"indianred": "#cd5c5c",
|
||||
"indigo": "#4b0082",
|
||||
"ivory": "#fffff0",
|
||||
"khaki": "#f0e68c",
|
||||
"lavender": "#e6e6fa",
|
||||
"lavenderblush": "#fff0f5",
|
||||
"lawngreen": "#7cfc00",
|
||||
"lemonchiffon": "#fffacd",
|
||||
"lightblue": "#add8e6",
|
||||
"lightcoral": "#f08080",
|
||||
"lightcyan": "#e0ffff",
|
||||
"lightgoldenrodyellow": "#fafad2",
|
||||
"lightgreen": "#90ee90",
|
||||
"lightgray": "#d3d3d3",
|
||||
"lightgrey": "#d3d3d3",
|
||||
"lightpink": "#ffb6c1",
|
||||
"lightsalmon": "#ffa07a",
|
||||
"lightseagreen": "#20b2aa",
|
||||
"lightskyblue": "#87cefa",
|
||||
"lightslategray": "#778899",
|
||||
"lightslategrey": "#778899",
|
||||
"lightsteelblue": "#b0c4de",
|
||||
"lightyellow": "#ffffe0",
|
||||
"lime": "#00ff00",
|
||||
"limegreen": "#32cd32",
|
||||
"linen": "#faf0e6",
|
||||
"magenta": "#ff00ff",
|
||||
"maroon": "#800000",
|
||||
"mediumaquamarine": "#66cdaa",
|
||||
"mediumblue": "#0000cd",
|
||||
"mediumorchid": "#ba55d3",
|
||||
"mediumpurple": "#9370db",
|
||||
"mediumseagreen": "#3cb371",
|
||||
"mediumslateblue": "#7b68ee",
|
||||
"mediumspringgreen": "#00fa9a",
|
||||
"mediumturquoise": "#48d1cc",
|
||||
"mediumvioletred": "#c71585",
|
||||
"midnightblue": "#191970",
|
||||
"mintcream": "#f5fffa",
|
||||
"mistyrose": "#ffe4e1",
|
||||
"moccasin": "#ffe4b5",
|
||||
"navajowhite": "#ffdead",
|
||||
"navy": "#000080",
|
||||
"oldlace": "#fdf5e6",
|
||||
"olive": "#808000",
|
||||
"olivedrab": "#6b8e23",
|
||||
"orange": "#ffa500",
|
||||
"orangered": "#ff4500",
|
||||
"orchid": "#da70d6",
|
||||
"palegoldenrod": "#eee8aa",
|
||||
"palegreen": "#98fb98",
|
||||
"paleturquoise": "#afeeee",
|
||||
"palevioletred": "#db7093",
|
||||
"papayawhip": "#ffefd5",
|
||||
"peachpuff": "#ffdab9",
|
||||
"peru": "#cd853f",
|
||||
"pink": "#ffc0cb",
|
||||
"plum": "#dda0dd",
|
||||
"powderblue": "#b0e0e6",
|
||||
"purple": "#800080",
|
||||
"rebeccapurple": "#663399",
|
||||
"red": "#ff0000",
|
||||
"rosybrown": "#bc8f8f",
|
||||
"royalblue": "#4169e1",
|
||||
"saddlebrown": "#8b4513",
|
||||
"salmon": "#fa8072",
|
||||
"sandybrown": "#f4a460",
|
||||
"seagreen": "#2e8b57",
|
||||
"seashell": "#fff5ee",
|
||||
"sienna": "#a0522d",
|
||||
"silver": "#c0c0c0",
|
||||
"skyblue": "#87ceeb",
|
||||
"slateblue": "#6a5acd",
|
||||
"slategray": "#708090",
|
||||
"slategrey": "#708090",
|
||||
"snow": "#fffafa",
|
||||
"springgreen": "#00ff7f",
|
||||
"steelblue": "#4682b4",
|
||||
"tan": "#d2b48c",
|
||||
"teal": "#008080",
|
||||
"thistle": "#d8bfd8",
|
||||
"tomato": "#ff6347",
|
||||
"turquoise": "#40e0d0",
|
||||
"violet": "#ee82ee",
|
||||
"wheat": "#f5deb3",
|
||||
"white": "#ffffff",
|
||||
"whitesmoke": "#f5f5f5",
|
||||
"yellow": "#ffff00",
|
||||
"yellowgreen": "#9acd32",
|
||||
}
|
||||
1218
testclass/lib/python3.12/site-packages/PIL/ImageDraw.py
Normal file
1218
testclass/lib/python3.12/site-packages/PIL/ImageDraw.py
Normal file
File diff suppressed because it is too large
Load Diff
243
testclass/lib/python3.12/site-packages/PIL/ImageDraw2.py
Normal file
243
testclass/lib/python3.12/site-packages/PIL/ImageDraw2.py
Normal file
@ -0,0 +1,243 @@
|
||||
#
|
||||
# The Python Imaging Library
|
||||
# $Id$
|
||||
#
|
||||
# WCK-style drawing interface operations
|
||||
#
|
||||
# History:
|
||||
# 2003-12-07 fl created
|
||||
# 2005-05-15 fl updated; added to PIL as ImageDraw2
|
||||
# 2005-05-15 fl added text support
|
||||
# 2005-05-20 fl added arc/chord/pieslice support
|
||||
#
|
||||
# Copyright (c) 2003-2005 by Secret Labs AB
|
||||
# Copyright (c) 2003-2005 by Fredrik Lundh
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
|
||||
|
||||
"""
|
||||
(Experimental) WCK-style drawing interface operations
|
||||
|
||||
.. seealso:: :py:mod:`PIL.ImageDraw`
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import Any, AnyStr, BinaryIO
|
||||
|
||||
from . import Image, ImageColor, ImageDraw, ImageFont, ImagePath
|
||||
from ._typing import Coords, StrOrBytesPath
|
||||
|
||||
|
||||
class Pen:
|
||||
"""Stores an outline color and width."""
|
||||
|
||||
def __init__(self, color: str, width: int = 1, opacity: int = 255) -> None:
|
||||
self.color = ImageColor.getrgb(color)
|
||||
self.width = width
|
||||
|
||||
|
||||
class Brush:
|
||||
"""Stores a fill color"""
|
||||
|
||||
def __init__(self, color: str, opacity: int = 255) -> None:
|
||||
self.color = ImageColor.getrgb(color)
|
||||
|
||||
|
||||
class Font:
|
||||
"""Stores a TrueType font and color"""
|
||||
|
||||
def __init__(
|
||||
self, color: str, file: StrOrBytesPath | BinaryIO, size: float = 12
|
||||
) -> None:
|
||||
# FIXME: add support for bitmap fonts
|
||||
self.color = ImageColor.getrgb(color)
|
||||
self.font = ImageFont.truetype(file, size)
|
||||
|
||||
|
||||
class Draw:
|
||||
"""
|
||||
(Experimental) WCK-style drawing interface
|
||||
"""
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
image: Image.Image | str,
|
||||
size: tuple[int, int] | list[int] | None = None,
|
||||
color: float | tuple[float, ...] | str | None = None,
|
||||
) -> None:
|
||||
if isinstance(image, str):
|
||||
if size is None:
|
||||
msg = "If image argument is mode string, size must be a list or tuple"
|
||||
raise ValueError(msg)
|
||||
image = Image.new(image, size, color)
|
||||
self.draw = ImageDraw.Draw(image)
|
||||
self.image = image
|
||||
self.transform: tuple[float, float, float, float, float, float] | None = None
|
||||
|
||||
def flush(self) -> Image.Image:
|
||||
return self.image
|
||||
|
||||
def render(
|
||||
self,
|
||||
op: str,
|
||||
xy: Coords,
|
||||
pen: Pen | Brush | None,
|
||||
brush: Brush | Pen | None = None,
|
||||
**kwargs: Any,
|
||||
) -> None:
|
||||
# handle color arguments
|
||||
outline = fill = None
|
||||
width = 1
|
||||
if isinstance(pen, Pen):
|
||||
outline = pen.color
|
||||
width = pen.width
|
||||
elif isinstance(brush, Pen):
|
||||
outline = brush.color
|
||||
width = brush.width
|
||||
if isinstance(brush, Brush):
|
||||
fill = brush.color
|
||||
elif isinstance(pen, Brush):
|
||||
fill = pen.color
|
||||
# handle transformation
|
||||
if self.transform:
|
||||
path = ImagePath.Path(xy)
|
||||
path.transform(self.transform)
|
||||
xy = path
|
||||
# render the item
|
||||
if op in ("arc", "line"):
|
||||
kwargs.setdefault("fill", outline)
|
||||
else:
|
||||
kwargs.setdefault("fill", fill)
|
||||
kwargs.setdefault("outline", outline)
|
||||
if op == "line":
|
||||
kwargs.setdefault("width", width)
|
||||
getattr(self.draw, op)(xy, **kwargs)
|
||||
|
||||
def settransform(self, offset: tuple[float, float]) -> None:
|
||||
"""Sets a transformation offset."""
|
||||
(xoffset, yoffset) = offset
|
||||
self.transform = (1, 0, xoffset, 0, 1, yoffset)
|
||||
|
||||
def arc(
|
||||
self,
|
||||
xy: Coords,
|
||||
pen: Pen | Brush | None,
|
||||
start: float,
|
||||
end: float,
|
||||
*options: Any,
|
||||
) -> None:
|
||||
"""
|
||||
Draws an arc (a portion of a circle outline) between the start and end
|
||||
angles, inside the given bounding box.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.arc`
|
||||
"""
|
||||
self.render("arc", xy, pen, *options, start=start, end=end)
|
||||
|
||||
def chord(
|
||||
self,
|
||||
xy: Coords,
|
||||
pen: Pen | Brush | None,
|
||||
start: float,
|
||||
end: float,
|
||||
*options: Any,
|
||||
) -> None:
|
||||
"""
|
||||
Same as :py:meth:`~PIL.ImageDraw2.Draw.arc`, but connects the end points
|
||||
with a straight line.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.chord`
|
||||
"""
|
||||
self.render("chord", xy, pen, *options, start=start, end=end)
|
||||
|
||||
def ellipse(self, xy: Coords, pen: Pen | Brush | None, *options: Any) -> None:
|
||||
"""
|
||||
Draws an ellipse inside the given bounding box.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.ellipse`
|
||||
"""
|
||||
self.render("ellipse", xy, pen, *options)
|
||||
|
||||
def line(self, xy: Coords, pen: Pen | Brush | None, *options: Any) -> None:
|
||||
"""
|
||||
Draws a line between the coordinates in the ``xy`` list.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.line`
|
||||
"""
|
||||
self.render("line", xy, pen, *options)
|
||||
|
||||
def pieslice(
|
||||
self,
|
||||
xy: Coords,
|
||||
pen: Pen | Brush | None,
|
||||
start: float,
|
||||
end: float,
|
||||
*options: Any,
|
||||
) -> None:
|
||||
"""
|
||||
Same as arc, but also draws straight lines between the end points and the
|
||||
center of the bounding box.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.pieslice`
|
||||
"""
|
||||
self.render("pieslice", xy, pen, *options, start=start, end=end)
|
||||
|
||||
def polygon(self, xy: Coords, pen: Pen | Brush | None, *options: Any) -> None:
|
||||
"""
|
||||
Draws a polygon.
|
||||
|
||||
The polygon outline consists of straight lines between the given
|
||||
coordinates, plus a straight line between the last and the first
|
||||
coordinate.
|
||||
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.polygon`
|
||||
"""
|
||||
self.render("polygon", xy, pen, *options)
|
||||
|
||||
def rectangle(self, xy: Coords, pen: Pen | Brush | None, *options: Any) -> None:
|
||||
"""
|
||||
Draws a rectangle.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.rectangle`
|
||||
"""
|
||||
self.render("rectangle", xy, pen, *options)
|
||||
|
||||
def text(self, xy: tuple[float, float], text: AnyStr, font: Font) -> None:
|
||||
"""
|
||||
Draws the string at the given position.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.text`
|
||||
"""
|
||||
if self.transform:
|
||||
path = ImagePath.Path(xy)
|
||||
path.transform(self.transform)
|
||||
xy = path
|
||||
self.draw.text(xy, text, font=font.font, fill=font.color)
|
||||
|
||||
def textbbox(
|
||||
self, xy: tuple[float, float], text: AnyStr, font: Font
|
||||
) -> tuple[float, float, float, float]:
|
||||
"""
|
||||
Returns bounding box (in pixels) of given text.
|
||||
|
||||
:return: ``(left, top, right, bottom)`` bounding box
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.textbbox`
|
||||
"""
|
||||
if self.transform:
|
||||
path = ImagePath.Path(xy)
|
||||
path.transform(self.transform)
|
||||
xy = path
|
||||
return self.draw.textbbox(xy, text, font=font.font)
|
||||
|
||||
def textlength(self, text: AnyStr, font: Font) -> float:
|
||||
"""
|
||||
Returns length (in pixels) of given text.
|
||||
This is the amount by which following text should be offset.
|
||||
|
||||
.. seealso:: :py:meth:`PIL.ImageDraw.ImageDraw.textlength`
|
||||
"""
|
||||
return self.draw.textlength(text, font=font.font)
|
||||
113
testclass/lib/python3.12/site-packages/PIL/ImageEnhance.py
Normal file
113
testclass/lib/python3.12/site-packages/PIL/ImageEnhance.py
Normal file
@ -0,0 +1,113 @@
|
||||
#
|
||||
# The Python Imaging Library.
|
||||
# $Id$
|
||||
#
|
||||
# image enhancement classes
|
||||
#
|
||||
# For a background, see "Image Processing By Interpolation and
|
||||
# Extrapolation", Paul Haeberli and Douglas Voorhies. Available
|
||||
# at http://www.graficaobscura.com/interp/index.html
|
||||
#
|
||||
# History:
|
||||
# 1996-03-23 fl Created
|
||||
# 2009-06-16 fl Fixed mean calculation
|
||||
#
|
||||
# Copyright (c) Secret Labs AB 1997.
|
||||
# Copyright (c) Fredrik Lundh 1996.
|
||||
#
|
||||
# See the README file for information on usage and redistribution.
|
||||
#
|
||||
from __future__ import annotations
|
||||
|
||||
from . import Image, ImageFilter, ImageStat
|
||||
|
||||
|
||||
class _Enhance:
|
||||
image: Image.Image
|
||||
degenerate: Image.Image
|
||||
|
||||
def enhance(self, factor: float) -> Image.Image:
|
||||
"""
|
||||
Returns an enhanced image.
|
||||
|
||||
:param factor: A floating point value controlling the enhancement.
|
||||
Factor 1.0 always returns a copy of the original image,
|
||||
lower factors mean less color (brightness, contrast,
|
||||
etc), and higher values more. There are no restrictions
|
||||
on this value.
|
||||
:rtype: :py:class:`~PIL.Image.Image`
|
||||
"""
|
||||
return Image.blend(self.degenerate, self.image, factor)
|
||||
|
||||
|
||||
class Color(_Enhance):
|
||||
"""Adjust image color balance.
|
||||
|
||||
This class can be used to adjust the colour balance of an image, in
|
||||
a manner similar to the controls on a colour TV set. An enhancement
|
||||
factor of 0.0 gives a black and white image. A factor of 1.0 gives
|
||||
the original image.
|
||||
"""
|
||||
|
||||
def __init__(self, image: Image.Image) -> None:
|
||||
self.image = image
|
||||
self.intermediate_mode = "L"
|
||||
if "A" in image.getbands():
|
||||
self.intermediate_mode = "LA"
|
||||
|
||||
if self.intermediate_mode != image.mode:
|
||||
image = image.convert(self.intermediate_mode).convert(image.mode)
|
||||
self.degenerate = image
|
||||
|
||||
|
||||
class Contrast(_Enhance):
|
||||
"""Adjust image contrast.
|
||||
|
||||
This class can be used to control the contrast of an image, similar
|
||||
to the contrast control on a TV set. An enhancement factor of 0.0
|
||||
gives a solid gray image. A factor of 1.0 gives the original image.
|
||||
"""
|
||||
|
||||
def __init__(self, image: Image.Image) -> None:
|
||||
self.image = image
|
||||
if image.mode != "L":
|
||||
image = image.convert("L")
|
||||
mean = int(ImageStat.Stat(image).mean[0] + 0.5)
|
||||
self.degenerate = Image.new("L", image.size, mean)
|
||||
if self.degenerate.mode != self.image.mode:
|
||||
self.degenerate = self.degenerate.convert(self.image.mode)
|
||||
|
||||
if "A" in self.image.getbands():
|
||||
self.degenerate.putalpha(self.image.getchannel("A"))
|
||||
|
||||
|
||||
class Brightness(_Enhance):
|
||||
"""Adjust image brightness.
|
||||
|
||||
This class can be used to control the brightness of an image. An
|
||||
enhancement factor of 0.0 gives a black image. A factor of 1.0 gives the
|
||||
original image.
|
||||
"""
|
||||
|
||||
def __init__(self, image: Image.Image) -> None:
|
||||
self.image = image
|
||||
self.degenerate = Image.new(image.mode, image.size, 0)
|
||||
|
||||
if "A" in image.getbands():
|
||||
self.degenerate.putalpha(image.getchannel("A"))
|
||||
|
||||
|
||||
class Sharpness(_Enhance):
|
||||
"""Adjust image sharpness.
|
||||
|
||||
This class can be used to adjust the sharpness of an image. An
|
||||
enhancement factor of 0.0 gives a blurred image, a factor of 1.0 gives the
|
||||
original image, and a factor of 2.0 gives a sharpened image.
|
||||
"""
|
||||
|
||||
def __init__(self, image: Image.Image) -> None:
|
||||
self.image = image
|
||||
self.degenerate = image.filter(ImageFilter.SMOOTH)
|
||||
|
||||
if "A" in image.getbands():
|
||||
self.degenerate.putalpha(image.getchannel("A"))
|
||||
Some files were not shown because too many files have changed in this diff Show More
Loading…
x
Reference in New Issue
Block a user