NaMAZU
ML utilities/models for PyTorch, PyTorch Lighting
Install / Use
/learn @NMZ0429/NaMAZUREADME
Many utilities for ML
</div>
NaMAZU
Installation
Version in pip server might be older than this repo.
pip install NaMAZU
1.Lightning API
1-1.Deep Learning Models
Collection of SOTA or robust baseline models for multiple tasks fully written in pytorch lightning! They are all ready-to-train models with MNIST, ImageNet, UCF101 etc... using LightingDataModule.
Some models come with their pretrained-weight available by auto-downloading.
import pytorch_lightning as pl
from NaMAZU.lightningwingman import LitVideoClf
config = {"num_classes": 10, "cnn": "resnet152d", "latent_dim":512}
model = LitVideoClf(use_lstm=False, model_config=config)
...
# use bolts to get datamodule and pass model and datamodule to pl.trainer!
- LitU2Net: LightningModule U2Net. Trainable and ready for prediction.
- AniNet: LightningModule image classifier pretrained for japanese animations.
- LitVideoClf: LightningModule video classfier using either single frame CNN or CNNLSTM.
- MultiModalNet: LightningModule for multi-modal learning which can learn any modality with high robustness. Can be combined with any backbone.
1-2.Feature Learning Interface
Before starting your fine-tuning training, try this trianign API that produces better initial weight by running a self-supervised learning to your training dataset. Only images are used and no annotation nor data cleaning is required.
Other training schemes are coming soon!
from NaMAZU.lightingwingman import self_supervised_learning
# images may be stored in single or multiple directories. Stratified sampling is supported!
dir_images = "dataset/something"
dir_images2 = "dataset/something2"
self_supervised_training(
"resnet50",
[dir_images, dir_images2],
batch_size=64,
save_dir="pretrained_models/"
)
- self_supervised_training: Simple interface that you can obtain self-supervised CNN with just one line of code!
1-3.Statistical Models
They are all written in PyTorch following best practice to be used with pytorch lightning. They are all GPU enabled controlled by Lightning API. You will never need to call to("cuda") to use the model on any device even with multi-GPU training!
import pytorch_lightning as pl
from NaMAZU.lightningwingman import KNN, GMM
class YourLitModule(pl.LightningModule):
def __init__(self,*args, **kwargs):
...
self.encoder = SomeEncoder()
self.head_classifier = KNN(
n_neighbors=5,
distance_measure="cosine",
training_data=some_known_data
)
self.estimator = GMM(5, 10)
def training_step(self, batch):
x, t = batch
y = self.encoder(x)
y_hat = self.head_classifier(y)
probability = self.estimator.predict_proba(y)
- KNN: Available with euqlidean, manhattan, cosine and mahalanobis distance.
- NBC: GPU enabled naive bayes classifier.
- GMM: Gaussian Mixture probabability estimator. Of course GPU enabled.
2.ONNX API
We provide many readly to use ONNX models comes with preprocess and postprocess methods. They are packed as an class object and you can use it without any coding!
Weight files are automatically downloaded to the currently working directory if you don't have it or you can load existing model.
- MiDAS: Mono Depth Prediction (Light and Large models are available)
- U2Net: Saliency Segmnentation (Available with 4 task-specified weights)
- RealESR: Super Resolution (3 models. Predict method directly return upscaled image)
from NaMAZU.onnxapi import MiDASInference
model = MiDASInference(model="mono_depth_large.onnx")
prediction = model.predict("some_image.jpg") # Accept cv2 image as well
result = model.render(prediction, "some_image.jpg")
plt.imshow(result)
3.Functional API
You can use below functions via
import NaMAZU.functional as F
F.change_frame_rates("./test_data.mp4",fps=5)
image_control
<details><summary>List of functions</summary><div>- npy_to_img
- img_to_npy
- split_image
- compose_two_png
- apply_mask_to
- apply_to_all
- change_frame_rates
- save_all_frames
- make_video_from_frames
- collect_images (requires icrawler)
file_control
<details><summary>List of functions</summary><div>- rename_file
- collect_file_pathes_by_ext
- zip_files
- randomly_choose_files
- export_list_str_as
text_control
<details><summary>List of functions</summary><div>- search_word_from
data_science [Under redesign]
<details><summary>List of functions</summary><div>- train_linear_regressor
- parse_tab_seperated_txt
Sampling Theory
- calculate_sample_stats
- error_bound_of_mean
- calculate_sufficient_n_for_mean
- estimated_total
- error_bound_of_total
- calculate_sufficient_n_for_population_total
- calculate_sufficient_n_for_proportion
- calculate_sufficient_n_for_proportion
Regression Analysis
- sxy_of
- sxx_of
- least_square_estimate
- estimate_variance_of_linear_regressor
- t_statistic_of_beta1
- calculate_CI_of_centred_model_at
- get_prediction_interval
- t_stats_for_correlation
- get_p_value_of_tstat
- fit_general_least_square_regression
Correlation Analysis
- "get_prediction_interval"
- "t_stats_for_correlation"
- "get_p_value_of_tstat"
- "_search_t_table"
- "get_alt_sxx"
- "get_alt_sxy"
coreml
<details><summary>List of functions</summary><div>- drop_negative
4.Visual Integration
st_utils
- hide_default_header_and_footer
- plot_plotly_supervised
5.Decorator [Under redesign]
Some utility decorators to speed up your development.
- print_docstring
- measure_runtime
:rocket: Coming
- [ ] 2. PredictionAssistant
- [x] 2. Video Recognition Model
- [x] 3. Feature Learning
- [ ] 4. Few-shot Learning
- [ ] 5. Audio-Visual Multimodal fusion (finish docstrings)
- [ ] 6. BBox template finding
- [ ] 7. CACNet
Related Skills
claude-opus-4-5-migration
109.7kMigrate prompts and code from Claude Sonnet 4.0, Sonnet 4.5, or Opus 4.1 to Opus 4.5
model-usage
349.7kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
TrendRadar
51.0k⭐AI-driven public opinion & trend monitor with multi-platform aggregation, RSS, and smart alerts.🎯 告别信息过载,你的 AI 舆情监控助手与热点筛选工具!聚合多平台热点 + RSS 订阅,支持关键词精准筛选。AI 智能筛选新闻 + AI 翻译 + AI 分析简报直推手机,也支持接入 MCP 架构,赋能 AI 自然语言对话分析、情感洞察与趋势预测等。支持 Docker ,数据本地/云端自持。集成微信/飞书/钉钉/Telegram/邮件/ntfy/bark/slack 等渠道智能推送。
mcp-for-beginners
15.8kThis open-source curriculum introduces the fundamentals of Model Context Protocol (MCP) through real-world, cross-language examples in .NET, Java, TypeScript, JavaScript, Rust and Python. Designed for developers, it focuses on practical techniques for building modular, scalable, and secure AI workflows from session setup to service orchestration.
