how to copy file, directory from PODs to local host machine.

how to copy file, directory from K8s POD to local machine ?

     Generally we need some data from K8s PODs to our local machine for debugging and for future references. In this blog we will see that how we can copy the data from K8s POD to the local host directory or we can say to our local host machine and vise versa.

Copy data from PODs to the Local machine. 

1- If we are in the POD.

       if you are in the K8s POD already, then you can use SCP commands and their options to send the data file/directory to the local machine. 

From below commands we can login into the PODs.

kubectl exec -it <POD-Name > /bin/bash

After that Go to that path where the file present. use below command to scp the file to the local machine.

scp -r <filename /directory name > root@<local-machine IP >:<local machine path>

for Ex:

scp -r dummy.pcap root@xx.xx.xx.xx:/home/localdirectory/


2- using kubectl cp command:

# Copy /tmp/foo from a remote pod to /tmp/bar locally

kubectl cp <some-namespace>/<some-pod>:/tmp/foo /tmp/bar

we can also use in below format to copy the folder from POD to local host.

kubectl cp server-cudu-f66785456-agbcp:apps/config  /home/name/files/config


Copy local directory to remote POD directory:

# Copy /tmp/foo local file to /tmp/bar in a remote pod in a specific container

kubectl cp /tmp/foo <some-pod>:/tmp/bar -c <specific-container>

# Copy /tmp/foo local file to /tmp/bar in a remote pod in namespace <some-namespace>

kubectl cp /tmp/foo <some-namespace>/<some-pod>:/tmp/bar

we can also use in below format to copy a folder from the localhost directory to a POD.

kubectl cp /home/name/files/config server-cudu-f56776c55646-abcde:apps/config


How to Check linux OS Version ?

How to Check CentOS Version ?

       There are several reasons, that why we need to check the centos version. For example if we need to downloads a SW package from the internet and install into the system. then we need to know that which version and OS bit is required to download. some times if we debugging some issue, generally we need check the ON version / SW version. 

so here we will see the multiple ways to check the CentOS version.

1- using rpm commands check Centos version:

rpm --query centos-release





2- using "hostnamectl" command:

hostnamectl command is supported by CentOS 7 and above versions. 

This command also display the Machine ID, Boot ID, hostname (server name) and kernel version also.

hostnamectl



3- using "cat /etc/os-release"

cat /etc/os-release



















For full OS version use command as below.

cat /etc/centos-release







4- How to check the system info.

      By using "lscpu" this command, we can check the number of Core, number of CPU, Number of NUMA node and NUMA setting, modal chipset, CPU Architecture and flags which are enabled in the systems.



5G-SA Call Flow: Messages mapped with channels:

 5G-SA Call Flow:

In this section of blog, we have tried to map some of the massages of 5G-SA call flow with channels (physical/Transport/Logical), SRBs over which messages are being transfer, mapped coreset and search space with the messages and RLC mode used during UE attached.




Install docker - Kubernetes on Ubuntu.

 

Steps to Install docker - Kubernetes on Ubuntu

Set up Docker

Step 1: Install Docker

Kubernetes requires an existing Docker installation on all nodes master and worker node. If you already have Docker installed, skip ahead to Step 2.

1. Update the package list with the command:

sudo apt-get update

2. Next, install Docker with the below command:

sudo apt-get install docker.io

3. after completing the docker instalation . Check the installation (and version) by entering the following:

sudo docker version



Step 2: Start and Enable Docker

1. Set Docker to launch at boot by entering the following:

sudo systemctl enable docker

2. Verify Docker is running:

sudo systemctl status docker

To start Docker if it’s not running:

sudo systemctl start docker



Install Kubernetes

Step 3: Add Kubernetes Signing Key(both master and worker node)

Since you are downloading Kubernetes from a non-standard repository, it is essential to ensure that the software is authentic. This is done by adding a signing key.

1. Enter the following to add a signing key:

curl -s https://packages.cloud.google.com/apt/doc/apt-key.gpg | sudo apt-key add

If you get an error that curl is not installed, install it with:

sudo apt-get install curl







Step 4: Add Software Repositories

Kubernetes is not included in the default repositories. To add them, enter the following:

sudo apt-add-repository "deb http://apt.kubernetes.io/ kubernetes-xenial main"

Repeat on each server node.

Step 5: Kubernetes Installation Tools

Kubeadm (Kubernetes Admin) is a tool that helps initialize a cluster. It fast-tracks setup by using community-sourced best practices. Kubelet is the work package, which runs on every node and starts containers. The tool gives you command-line access to clusters.

1. Install Kubernetes tools with the command:

sudo apt-get install kubeadm kubelet kubectl
sudo apt-mark hold kubeadm kubelet kubectl

Allow the process to complete.

2. Verify the installation with:

kubeadm version

3. Repeat for each server node.



Kubernetes Deployment

Step 6: Begin Kubernetes Deployment

Start by disabling the swap memory on each server:

sudo swapoff –a

If any issue as below while executing above commsnd:

[ERROR Swap]: running with swap on is not supported. Please disable swap.

1- sudo kubeadm reset

2- Create a file in /etc/systemd/system/kubelet.service.d/20-allow-swap.conf with the content:

[Service]
Environment="KUBELET_EXTRA_ARGS=--fail-swap-on=false"

3- sudo swapoff –a



Step 7: Assign Unique Host name for Each Server Node 

Decide which server to set as the master node. Then enter the command:

sudo hostnamectl set-hostname master-node

Next, set a worker node hostname by entering the following on the worker server:

sudo hostnamectl set-hostname worker01

If you have additional worker nodes, use this process to set a unique hostname on each. For example:

worker node 1:

sudo hostnamectl set-hostname worker01

worker node 2:

sudo hostnamectl set-hostname worker02

Step 8: Initialize Kubernetes on Master Node only

Switch to the master server node, and enter the following:

sudo kubeadm init --pod-network-cidr=192.168.0.0/16

Once this command finishes, it will display a kubeadm join message at the end. Make a note of the whole entry. This will be used to join the worker nodes to the cluster.



Next, enter the following to create a directory for the cluster:

mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config



Step 9: Install Calico

  1. Install the Tigera Calico operator and custom resource definitions.

    sudo kubectl create -f https://docs.projectcalico.org/manifests/tigera-operator.yaml
  2. Install Calico by creating the necessary custom resource.

    sudo kubectl create -f https://docs.projectcalico.org/manifests/custom-resources.yaml
    Note: Before creating this manifest, read its contents and make sure its settings are correct for your environment. For example, you may need to change the default IP pool CIDR to match your pod network CIDR.
  3. Confirm that all of the pods are running with the following command.

    watch kubectl get pods -n calico-system

    Wait until each pod has the STATUS of Running.

  4. Confirm that you now have a node in your cluster with the following command.

      kubectl get nodes -o wide

Step 10: Join Worker Node to Cluster

As indicated in Step 7, you can enter the kubeadm join command on each worker node to connect it to the cluster.

1- Switch to the worker01 system and enter the command you noted from Step 7:

Repeat for each worker node on the cluster. Wait a few minutes; then you can check the status of the nodes.

2- Switch to the master server, and enter:

kubectl get nodes

The system should display the worker nodes that you joined to the cluster.



Step-11: install registry

Use a command like the following to start the registry container:

sudo docker run -d -p 5000:5000 --restart=always --name registry registry:2

note: If error regarding registry already exsits. Remove the registry by below command and then recreate.

sudo docker container stop registry && sudo docker container rm -v registry

After that check the status of socker, below hiligeted line will added in status.

sudo systemctl status docker



Test with insecure registry

This procedure configures Docker to entirely disregard security for your registry. This is very insecure and is not recommended. It exposes your registry to trivial man-in-the-middle (MITM) attacks. Only use this solution for isolated testing or in a tightly controlled, air-gapped environment.

  1. Edit the daemon.json file, whose default location is /etc/docker/daemon.json on Linux

  2. If the daemon.json file does not exist, create it. Assuming there are no other settings in the file, it should have the following contents:

{
  "insecure-registries" : ["myregistrydomain.com:5000"]
}



3. Restart Docker for the changes to take effect.

Repeat steps-11 on every Engine host that wants to access your registry.



Step-12 – Intall helm

-From Apt (Debian/Ubuntu)

Members of the Helm community have contributed a helm package for Apt. This package is generally up to date.

curl https://baltocdn.com/helm/signing.asc | sudo apt-key add -
sudo apt-get install apt-transport-https --yes
echo "deb https://baltocdn.com/helm/stable/debian/ all main" | sudo tee /etc/apt/sources.list.d/helm-stable-debian.list
sudo apt-get update
sudo apt-get install helm









kubernetes show deployment and delete deployment

 

show deployment

$ kubectl get deployments;

NAME              DESIRED   CURRENT   UP-TO-DATE   AVAILABLE   AGE
spring-hello      1         1         1            1           22h
spring-world      1         1         1            1           22h
vfe-hello-wrold   1         1         1            1           14m

show services

$kubectl get services;

NAME              TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)          AGE
kubernetes        ClusterIP   10.96.0.1        <none>        443/TCP          2d
spring-hello      NodePort    10.103.27.226   <none>        8081:30812/TCP   23h
spring-world      NodePort    10.102.21.165    <none>        8082:31557/TCP   23h
vfe-hello-wrold   NodePort    10.101.23.36     <none>        8083:31532/TCP   14m

delete deployment

$ kubectl delete deployments vfe-hello-wrold

deployment.extensions "vfe-hello-wrold" deleted

delete services

$ kubectl delete service vfe-hello-wrold

service "vfe-hello-wrold" deleted


5G-NR: NRF


Role of NRF in 5G architecture:

Network Repository Function (NRF)

The NRF maintains a record of all the 5G elements that are available in the network and their supported services. It allows other NF instances to subscribe and be notified of registrations from NF instances of a given type


The NRF supports discovery mechanisms that allows 5G elements to discover each other and get updated status of the desired elements.

The NRF supports the following functions:

  • Maintains the profiles of the available NF instances and their supported services in the 5G core network
  • Allows consumer NF instances to discover other providers NF instances in the 5G core network
  • Allows NF instances to track the status of other NF instances

 It supports the above functions through the following services:

  • Management Services (Nnrf_NFManagement)
  • Discovery Services (Nnrf_NFDiscovery)

Management Services (nrf_NFManagement.

This will handles the following service operations:
  • Receives and handles NFRegister service requests from the Nfs.
  • Receives and handles NFDeregister service requests from the NFs
  • Stores the registered profiles in its own data store using the database service.
  • Receives and handles NFDeregister service requests from the Nfs.
  • Receives and handles the Heart-beat messages from the NFs
  • Monitors the heart-beat expiry, mark the NF profiles as suspended and take appropriate action on the suspended NF profiles.
  • NF profile Retrieval.
  • Receives and handles NFStatusSubscribe service requests from the NFs.
  • Receives and handles NFStatusUnsubscribe service requests from the NFs.
  • Receives and handles NFNotify service requests from the NFs
  • Stores the subscription data in its own data store using the database service
Discovery Services (Nnrf_NFDiscoveryService)

This will handles the following service operations:
  • Receives and handles NFDiscover service requests from the Nfs.

 
NRF Functions/Procedure Support:

The NRF interacts with every other element in the 5G core network and it supports the above functions through the following services over HTTPS protocols:
  • Management Services
  • Discovery Services


Supports the following Functions/Procedure-
1) NRF Management Services-
The NRF Management service is identified by the service operation name Nnrf_NFManagement. NRF supports the following management services.
 
Register NF instance (NFRegister):- It allows an NF Instance to register its NF profile in the NRF; it includes the registration of the general parameters of the NF Instance, together with the list of services exposed by the NF Instance. This service operation is not allowed to be invoked from an NRF in a different PLMN. 
 
Update NF instance (NFUpdate): Enables an NF instance to partially update or replace the parameters of its NF profile in the NRF. It also allows to add or delete services provided by the NF instance.
 
De-register NF instance (NFDeregister): It allows an NF Instance to de-register its NF profile in the NRF, including the services offered by the NF Instance. This service operation is not allowed to be invoked from an NRF in a different PLMN.
 
Subscribe to Status (NFStatusSubscribe): Enables an NF instance to subscribe the status changes of other NF instances registered in the NRF.
 
Unsubscribe to Status (NFStatusUnsubscribe): Enables an NF instance to unsubscribe the status changes of other NF instances.
 
Receive Notifications of Status (NFStatusNotify): Enables the NRF to notify changes status of NF instances to any subscriber of NF status. Changes also include information regarding newly registered and de-registered NFs.


2) Discovery Service:

The NRF Discovery service is identified by the service operation name Nnrf_NFDiscoveryService.
Nnrf_NFDiscoveryService- It also allows an NF to subscribe to be notified of registration, de-registration and profile changes of NF Instance along with their NF Services.