Showing posts with label realtime. Show all posts
Showing posts with label realtime. Show all posts

February 17, 2023

How to password pass while connect ssh in the command line

Now a days, it's very difficult remember the vm passwords while connecting in the office network. So, Connect easily using sshpass.

For installation for mac/Linux:

Mac:

brew install hudochenkov/sshpass/sshpass

Linux:

yum install sshpass


Example of usage:

Ex:

sshpass -p test123 ssh root@192.168.1.1


Reafference:

https://stackoverflow.com/questions/32255660/how-to-install-sshpass-on-mac 

Read more ...

November 16, 2022

SRIOV CNI Plugin

The Single Root I/O Virtualization (SR-IOV) specification is a standard for a type of PCI device assignment that can share a single device with multiple pods. 

SR-IOV enables you to segment a compliant network device, recognized on the host node as a physical function (PF), into multiple virtual functions (VFs), and make them available for direct IO to the POD.

This plugin enables the configuration and usage of SR-IOV VF networks in containers and orchestrators like Kubernetes. 

Network Interface Cards (NICs) with SR-IOV capabilities are managed through physical functions (PFs) and virtual functions (VFs). A PF is used by the host and usually represents a single NIC port. VF configurations are applied through the PF. With SR-IOV CNI each VF can be treated as a separate network interface, assigned to a container, and configured with it's own MAC, VLAN, IP and more.

SR-IOV CNI plugin works with SR-IOV device plugin for VF allocation in Kubernetes. A metaplugin such as Multus gets the allocated VF's deviceID(PCI address) and is responsible for invoking the SR-IOV CNI plugin with that deviceID.

The end result will be similar to the in the picture except for the SRIOV-CNI and the DPDK userspace.


Reference:

https://github.com/ramanujadasu/sriov-cni

https://dramasamy.medium.com/high-performance-containerized-applications-in-kubernetes-f494cef3f8e8

Read more ...

Understanding the Kubernetes Node

Kubernetes is an open-source orchestration engine for automating deployments, scaling, managing, and providing the infrastructure to host containerized applications. At the infrastructure level, a Kubernetes cluster is comprised of a set of physical or virtual machines, each acting in a specific role.

Master components are responsible for managing the Kubernetes cluster. They manage the life cycle of pods, the base unit of a deployment within a Kubernetes cluster. Master servers run the following components:

kube-apiserver – the main component, exposing APIs for the other master components.

etcd – distributed key/value store which Kubernetes uses for persistent storage of all cluster information.

kube-scheduler – uses information in the pod spec to decide on which node to run a pod.

kube-controller-manager – responsible for node management (detecting if a node fails), pod replication, and endpoint creation.

cloud-controller-manager – daemon acting like an abstraction layer between the APIs and the different cloud providers’ tools (storage volumes, load balancers etc.)


Node components are worker machines in Kubernetes and are managed by the Master. A node may be a virtual machine (VM) or physical machine, and Kubernetes runs equally well on both types of systems. Each node contains the necessary components to run pods:

kubelet – watches the API server for pods on that node and makes sure they are running

cAdvisor – collects metrics about pods running on that particular node

kube-proxy – watches the API server for pods/services changes in order to maintain the network up to date

container runtime – responsible for managing container images and running containers on that node


Reference:

https://www.suse.com/c/rancher_blog/understanding-the-kubernetes-node/#:~:text=kubelet%20%E2%80%93%20watches%20the%20API%20server,the%20network%20up%20to%20date

Read more ...

April 26, 2022

Create New VPN entry in Cisco AnyConnect

MacBook process to add new VPN entry:


cd /opt/cisco/anyconnect/profile

sudo cp clientprofile_old.xml clientprofile_new.xml

sudo vi clientprofile_new.xml

    Update hostname and hostaddress

    Ex:

        <ServerList>

                <HostEntry>

                        <HostName>NEW</HostName>

                        <HostAddress>127.0.0.1</HostAddress>

                </HostEntry>

         </ServerList>

 

Few mac tools for easy connect Virtual Machines:

https://mobaxterm.mobatek.net/

https://apps.apple.com/us/app/zen-term-lite-ssh-client/id1422475219?mt=12

https://iterm2.com/

 

Reference:

https://documentation.meraki.com/MX/AnyConnect_on_the_MX_Appliance/Client_deployment

Read more ...

April 15, 2022

DevOps vs GitOps

DevOps is about cultural change and providing a way for development teams and operations teams to work together collaboratively. GitOps gives you tools and a framework to take DevOps practices, like collaboration, CI/CD, and version control, and apply them to infrastructure automation and application deployment.

GitOps Free Training(LinuxFoundation):

https://trainingportal.linuxfoundation.org/learn/course/introduction-to-gitops-lfs169/course-introduction/course-information 

Reference:

https://www.redhat.com/en/topics/devops/what-is-gitops#:~:text=DevOps%20is%20about%20cultural%20change,infrastructure%20automation%20and%20application%20deployment.

Read more ...

February 16, 2022

Pushing Images to artifactory or any location of docker environment

Skopeo is a tool for moving container images between different types of container storages. It allows you to copy container images between container registries like docker.io, quay.io, and your internal container registry or different types of storage on your local system.

Below are the steps to push or pull images from artifactory using skopeo instead of docker. Skopeo is just a cli and does not require any service/privilege. No change is needed on the docker or docker-desktop settings also.


Pre-Steps: Install skopeo

brew instal skopeo

Setup proxies if any issue with environment:

export http_proxy=<hostname>:<port>

export https_proxy=<hostname>:<port>

export no_proxy="io,com,org"

Login:

skopeo login --tls-verify=false -u USERNAME <repo-url>

Inspecting a repository

skopeo inspect --config docker://quay.io/podman/stable | json_pp

Copying images

Skopeo can copy container images between various storage mechanisms:

skopeo copy docker://<imagename:latest> docker://<imagename:latest>

ex: skopeo copy docker://registry.access.redhat.com/ubi8-init docker://reg.company.com/ubi-init

skopeo copy docker://registry.fedoraproject.org/fedora:latest  containers-storage:fedora

If any restrictions repo: skopeo copy --override-arch=amd64 --override-os=linux --dest-tls-verify=false docker-daemon:<imagename:latest> docker://<imagename:latest>

Deleting images from a registry:

skopeo delete docker://localhost:5000/<imagename:latest>


Reference:

https://www.redhat.com/en/blog/skopeo-10-released#:~:text=Skopeo%20is%20a%20tool%20for,storage%20on%20your%20local%20system

Read more ...

August 23, 2021

Python sWSGI.ini applications configurations

A note on Python threads

If you start uWSGI without threads, the Python GIL will not be enabled, so threads generated by your application will never run. You may not like that choice, but remember that uWSGI is a language-independent server, so most of its choices are for maintaining it “agnostic”. But do not worry, there are basically no choices made by the uWSGI developers that cannot be changed with an option.


If you want to maintain Python threads support without starting multiple threads for your application, just add the --enable-threads option (or enable-threads = true in ini style).


Ex:
 
[uwsgi]
# -------------
# Settings:
# key = value
# Comments >> #
# -------------

strict
uid = root
gid = root

# socket = [addr:port]
http-socket = 0.0.0.0:9034

wsgi-file = app.py
callable = TEST_SERVICE

# master = [master process (true of false)]
master = true

# processes = [number of processes]
processes = 5

py-autoreload = 0

buffer-size = 32768

need-app = true
attach-daemon = celery -A app.CELERY worker --loglevel=INFO --concurrency=1

Above properties detail information available below url:

Reference:

https://uwsgi-docs.readthedocs.io/en/latest/WSGIquickstart.html


Read more ...

My Favorite Site's List

#update below script more than 500 posts