116 skills found · Page 2 of 4
Murtaza-Saeed / Dlib Precompiled Wheels For Python On Windows X64 Easy InstallationDlib Compiled Binary Wheels for Python 3.7 - 3.12 on Windows (x64 CPUs)
cryybash / VoidVoid is a terminal based text editor written in Python, it is on a mission to push the boundaries of what traditional terminal editors never attempt. This editor is not trying to reinvent the wheel, it is merely trying to fill the unexplored space between terminal and GUI based editors.
mudigosa / Image ClassifierImage Classifier Going forward, AI algorithms will be incorporated into more and more everyday applications. For example, you might want to include an image classifier in a smartphone app. To do this, you'd use a deep learning model trained on hundreds of thousands of images as part of the overall application architecture. A large part of software development in the future will be using these types of models as common parts of applications. In this project, you'll train an image classifier to recognize different species of flowers. You can imagine using something like this in a phone app that tells you the name of the flower your camera is looking at. In practice, you'd train this classifier, then export it for use in your application. We'll be using this dataset of 102 flower categories. When you've completed this project, you'll have an application that can be trained on any set of labelled images. Here your network will be learning about flowers and end up as a command line application. But, what you do with your new skills depends on your imagination and effort in building a dataset. This is the final Project of the Udacity AI with Python Nanodegree Prerequisites The Code is written in Python 3.6.5 . If you don't have Python installed you can find it here. If you are using a lower version of Python you can upgrade using the pip package, ensuring you have the latest version of pip. To install pip run in the command Line python -m ensurepip -- default-pip to upgrade it python -m pip install -- upgrade pip setuptools wheel to upgrade Python pip install python -- upgrade Additional Packages that are required are: Numpy, Pandas, MatplotLib, Pytorch, PIL and json. You can donwload them using pip pip install numpy pandas matplotlib pil or conda conda install numpy pandas matplotlib pil In order to intall Pytorch head over to the Pytorch site select your specs and follow the instructions given. Viewing the Jyputer Notebook In order to better view and work on the jupyter Notebook I encourage you to use nbviewer . You can simply copy and paste the link to this website and you will be able to edit it without any problem. Alternatively you can clone the repository using git clone https://github.com/fotisk07/Image-Classifier/ then in the command Line type, after you have downloaded jupyter notebook type jupyter notebook locate the notebook and run it. Command Line Application Train a new network on a data set with train.py Basic Usage : python train.py data_directory Prints out current epoch, training loss, validation loss, and validation accuracy as the netowrk trains Options: Set direcotry to save checkpoints: python train.py data_dor --save_dir save_directory Choose arcitecture (alexnet, densenet121 or vgg16 available): pytnon train.py data_dir --arch "vgg16" Set hyperparameters: python train.py data_dir --learning_rate 0.001 --hidden_layer1 120 --epochs 20 Use GPU for training: python train.py data_dir --gpu gpu Predict flower name from an image with predict.py along with the probability of that name. That is you'll pass in a single image /path/to/image and return the flower name and class probability Basic usage: python predict.py /path/to/image checkpoint Options: Return top K most likely classes: python predict.py input checkpoint ---top_k 3 Use a mapping of categories to real names: python predict.py input checkpoint --category_names cat_To_name.json Use GPU for inference: python predict.py input checkpoint --gpu Json file In order for the network to print out the name of the flower a .json file is required. If you aren't familiar with json you can find information here. By using a .json file the data can be sorted into folders with numbers and those numbers will correspond to specific names specified in the .json file. Data and the json file The data used specifically for this assignemnt are a flower database are not provided in the repository as it's larger than what github allows. Nevertheless, feel free to create your own databases and train the model on them to use with your own projects. The structure of your data should be the following: The data need to comprised of 3 folders, test, train and validate. Generally the proportions should be 70% training 10% validate and 20% test. Inside the train, test and validate folders there should be folders bearing a specific number which corresponds to a specific category, clarified in the json file. For example if we have the image a.jpj and it is a rose it could be in a path like this /test/5/a.jpg and json file would be like this {...5:"rose",...}. Make sure to include a lot of photos of your catagories (more than 10) with different angles and different lighting conditions in order for the network to generalize better. GPU As the network makes use of a sophisticated deep convolutional neural network the training process is impossible to be done by a common laptop. In order to train your models to your local machine you have three options Cuda -- If you have an NVIDIA GPU then you can install CUDA from here. With Cuda you will be able to train your model however the process will still be time consuming Cloud Services -- There are many paid cloud services that let you train your models like AWS or Google Cloud Coogle Colab -- Google Colab gives you free access to a tesla K80 GPU for 12 hours at a time. Once 12 hours have ellapsed you can just reload and continue! The only limitation is that you have to upload the data to Google Drive and if the dataset is massive you may run out of space. However, once a model is trained then a normal CPU can be used for the predict.py file and you will have an answer within some seconds. Hyperparameters As you can see you have a wide selection of hyperparameters available and you can get even more by making small modifications to the code. Thus it may seem overly complicated to choose the right ones especially if the training needs at least 15 minutes to be completed. So here are some hints: By increasing the number of epochs the accuracy of the network on the training set gets better and better however be careful because if you pick a large number of epochs the network won't generalize well, that is to say it will have high accuracy on the training image and low accuracy on the test images. Eg: training for 12 epochs training accuracy: 85% Test accuracy: 82%. Training for 30 epochs training accuracy 95% test accuracy 50%. A big learning rate guarantees that the network will converge fast to a small error but it will constantly overshot A small learning rate guarantees that the network will reach greater accuracies but the learning process will take longer Densenet121 works best for images but the training process takes significantly longer than alexnet or vgg16 *My settings were lr=0.001, dropoup=0.5, epochs= 15 and my test accuracy was 86% with densenet121 as my feature extraction model. Pre-Trained Network The checkpoint.pth file contains the information of a network trained to recognise 102 different species of flowers. I has been trained with specific hyperparameters thus if you don't set them right the network will fail. In order to have a prediction for an image located in the path /path/to/image using my pretrained model you can simply type python predict.py /path/to/image checkpoint.pth Contributing Please read CONTRIBUTING.md for the process for submitting pull requests. Authors Shanmukha Mudigonda - Initial work Udacity - Final Project of the AI with Python Nanodegree
martin-riedl / CALexaCALexa is an Amazon Alexa Skill which allows accessing CalDAV Calendars. It makes use of the Flask-Ask Python micro-framework developed by John Wheeler.
ycheng517 / Ros Python WheelsSelf-contained Python wheels of ROS packages
repo-helper / WheyA simple Python wheel builder for simple projects.
dheera / Python FordreaderPython library to read steering wheel angles and more from a Ford car over OBD
aboutcode-org / Python Publicsuffix2A small Python library to deal with publicsuffix data (includes a bundled PSL as "package data") in a wheel friendly format. Fork and continuation of Tomaž Šolc's "publicsuffix"
dongfangliu / NH ORCA PythonThis is a pybind11 version of rvo2. Based on the python wrapper, an NH-ORCA algorithm is implemented for two-wheeled robots (such as turtle bot).
py-actions / Py Dependency InstallA GitHub Action that installs Python package dependencies from a user-defined requirements.txt file path with optional pip, setuptools, and wheel installs/updates
torchsmoke / Python3 WheelsWheels for Python 3
MrMino / Wheelfile🔪🧀 API for creating and inspecting Python .whl files (wheels)
Nishant-Wadhwani / Intelligent Infotainment System For Automobiles# Intellifotainment assist” – Smart HMI for passenger cars To run the program, download all files and save them in the same directory. After that, simply run 'Master.py' in the terminal. At the moment, the program will only run in linux based systems. # The Idea Infotainment systems have come a long way since the first set of dashboards installed in cars. Through our idea, we aim to create a Human Machine interaction model that takes infotainment systems to a new level. The driver tends to get distracted from the road while performing secondary tasks such as changing the music track, locking/unlocking the door while driving etc. Our system shall enable the driver to focus only on driving. Controlling the secondary tasks will be much easier. Our product primarily comprises of 5 modules: 1) Attention and drowsiness detection: - A camera shall be present on the dashboard, in front of the driver, behind the steering wheel. Through digital image processing techniques , using hough circle algorithm and haarcascade of an eye, we shall keep track of the driver’s sight. If he or she is looking away from the road while driving for more than a specified amount of time, we shall alert the driver to focus. We shall map the head orientation and iris position to accurately identify the driver’s attention. 2) Infotainment control features using blink combination: Through a combination of blinks, the driver can turn on or of the headlights, tail lights as well as indicators. Blinking of the eyes shall be detected using ‘dlib’ features in python. This shall give extremely accurate results. 3) Voice commands to control wipers, car lock, music system and windows A simple, yet extremely useful idea that would make the life of the driver a whole lot easy. Enabling the driver to speak to his car infotainment system would allow him to control and navigate these functionalities with great ease. The car will be enabled with a virtual assistant. 4) Automatic rear view mirror adjustment scheme: Using the camera placed in front of the driver, the system shall detect the position of the driver’s head. This shall also be done using image processing techniques and we shall identify the coordinates of the driver head in 3D space. There will be a mapping between the head position and mirror adjustment scheme. The mirrors will adjust their position using servo motors and shall do so automatically by identifying the head position. 5) A revolutionary reverse-assistance algorithm for smart parking and general reversing: Probably the highlight of our model, this feature shall make driving the car in and out of a parking spot, or rather, even reversing a car in general, far easier and safer than what it already is. Like most other modern cars, our model shall also have a camera installed at the back and the corresponding image displayed at the infotainment screen for parking assistance. Upon activating the reverse gear, the screen shall trace the line of motion of the car corresponding to the current position of the steering wheel. Because of this feature, the driver gets an idea of whether or not he’ll hit an obstacle while reversing if the steering wheel is kept at that position. Taking this feature to another level, the rear camera, after capturing the live video feed from the back of the car, shall perform image processing and machine learning algorithms to find a safe, obstacle-free path for reversing and indicate the driver to move the steering wheel accordingly. So instead of relying solely on the drivers judgement, our system shall actually find the path to be taken while reversing, such that other cars and other obstacles will be avoided, and accordingly recommend the driver to steer the wheel in that direction. This feature shall be extremely useful for new drivers/ learners. During the initial phase, to prevent errors from creeping in, we will always have a manual override button. After a good amount of testing, further modifications and refinements can be made. Our systems adds new dimensions to both precautionary safety measures, as well as convenience. If implemented properly, we are confident that our project will reach new heights of HMI and driver assistance technology. It will give drivers several less reasons to worry about.
nightlark / Swig Pypipip install swig. SWIG Python wheel for PyPI.
swag-engineering / Simbind CliGenerate Python wheel package from Simulink model!
Davidelanz / Soccer Match RobotFour-wheeled robot interfaced with ROS through a Raspberry Pi. Low level control of the L298N motor drivers via Arduino UNO controller. Developed ROS modules for perception tasks, visual servoing, and task planning in C++ and Python.
mrjadkowski / SMARS Mini XLA small wheeled or tracked robot running on Circuit Python
ameli / Manylinux Cudamanylinux docker images with CUDA Toolkit
boneylizard / Llama Cpp Python Cu128 Gemma3CUDA 12.8 accelerated prebuilt wheel of llama-cpp-python with full Gemma 3 model support for Windows 10/11 (x64). Built by boneylizard.
SagarBiswas-MultiHAT / SafeTodoManagerSafeTodoManager — an offline‑first, production‑ready Python To‑Do manager with a polished Tkinter UI, keyboard shortcuts, and flexible storage backends (filesystem, SQLite, optional encrypted files). Provides atomic saves, rotating backups/logs, headless CLI for export/backup, packaging (wheel/PyInstaller), and complete dev/test tooling.