Sei sulla pagina 1di 8

CORE NETWORK Core or central portion of a telecommunication network that is necessary for providing all the services to customer

that is connected to the access network is known as core network. Some of the main functions of the core network are to route or switch call across the PSTN and to provide path for the exchange of information between the different sub-networks. NETWORK TOPOLOGY It is the design or layout pattern of all the interconnections of various elements of a network or core network. It could be designed in both physical and logical pattern. Cable installation location and devices used refers to the physical topology, while the logical topology is about how data is transferred actually. Shape or structure of a network can be understood by the help of topology. Design Proposals for the core network based on lowest cost. For providing reliable connection for the transmission there should be use of such a topology which provides continuous data transmission. As here our area is not concern in a small region so it would be better if we use mixed network topology in which LAN & WLAN both are in use. As our production is distributed with in a building in between two floors so here we may use LAN as LAN itself having high data transmission rate. Using LAN for a single building also looks quite cost efficient too. Use of LAN will helps in data transfer up to a great transfer rate of in GBPS as due to being connected with wire transfer of data will be there with a huge transfer rate. While in case of WLAN wireless network transfer rate will be there only in MBPS because wireless network are quite slow as compare to wired one and secondly as in wireless networks there is no such data transferring media is there. In case of remote data transmission like between two buildings we may use mixed topology. In mixed topology WLAN is used to making connection between two buildings and afterwards within that building also LAN should be used. By using such a network topology SMG servers will be available for both live and recording feeds. Due to use of mixed topology first it will be quite cost efficient and secondly the maintenance of such network topology will be quite easy to handle and ease in keeping record of all also. There may be in somewhere small depth in data transfer rate due to change in channel mode

from LAN to WLAN and vice-versa. But if possible it will be better if we replace the whole on LAN itself. For end-user connection at the time of events like concerts and sports servers should be in high-availability condition, and should be able to handle the upcoming traffic. As concerts and sports both are required to be online all the time as unavailability of service may result in loose of revenue and customers confidence for service. End-users only want that the content for which they are looking for must be precise and do not let them to wait for a long time for accessing it. So the main target to attract the traffic in our content is to make end-user comfortable enough for their contents and make those contents highly available for them 24x7.

Design proposal on the basis of Best performance: High performance networks are very much essential and recently addition to tools used by plans and purchasers in order to attempt for cost reduction. In terms of efficiency and quality improvement these are often in description. These networks are in prevalence and are introduces only in selected markets, and in these markets these are offered without any type of enrolment and as they are distributed without any type of enrolment these are largely dependent on employers which includes high-performance networks as their sponsored network infrastructure. Network with high-performance are generally not distinct in their products, but they are just an option for different platform dependent products and most commonly they are preferred provider organization (PPOs). The high-performance networks differ across their plans as per their exact specifications. The cost-sharing differentials are corresponds the most commonly model uses tiered-provider levels. The first tier consists of the highperforming providers and secondly the second tier may consist of remainder of in-network providers and the last one the third consists of out-of-network providers. The employers often do not differentiate cost sharing with the first and the second tiers as the first having high-performing while the second consists of remainder of in-network providers, offering these networks only as a source of information to their employee abut which providers have better performance.

The most often target physicians plan are generally specialists for high performance networks. Multispecialty groups may practice pressed plans to include all of their specialities as assessments of companies to include in these networks are conducted by their speciality. And these plan criterias are used to select specialties to include focus on these following things: Representation of Information sharing on a very large scale. Reflecting variation significance of cost and quality up to great extent. With the practice level efficiency and quality there is generation of sufficient claim to the clients. Benchmark performances have been established on the basis of quality measure and/or guidelines.

Data storage capacity for sports videos: After recording of each and every concert or sports events SMG needs to store that recording for the website. And for that they need a huge data storage and that data should also be accessible online as end-users needs to view these recordings as per there requirements. So, for accessing such data through huge traffic these data should be mirrored to more than one link. Mirroring of data may help in handling web traffic for these recording on website. With the mirroring a single websites gets open from different servers which increases the website accessing speed and also increase the access rating of that website. Enduser do not get to feel any type of difference during accessing the mirrored website and feels satisfying from the service. And the only things to increase the traffic is to make the end user feel satisfying by the service. For storage RAID architecture may be used in which data will be stored at more than one place and it will also help in web accessibility of this data. By using RAID data gets mirrored at more than one place or we can say same data gets stored at more than one place with same type of storage structure. Benefits of doing so is that we are removing the risk of any type of disaster also as due to mirroring we not only copying the data at many places but also developing such a system which make us capable to recover the whole data if it get damaged any how during any miss happenings. By the use of RAID data is also become safe and easy to recover in case of any disaster. As with the help of any of the mirrored data the whole data set is easily

recoverable and reusable too. In case of resilience also there is no need to transfer the whole setup to any other place but just move any mirror and recovery that whole data there. Each node of mirror will go to keep a keyset which will be used to recover the data set. With the help of certain data set and the keyset RAID mirror image can be able to recover the whole data set related to that key and that datasets. So, here by the name of one we can do two jobs one is to handle the website traffic by diverting the traffic to mirrored sites and making the clients to access the data from different positions for the users from different regions. And secondly we here by preparing ourselves for disaster also in such a way that anyhow by any sort of damage we will be able to recover our whole data as such.

Minimisation of Server Computer Architecture investments: Traffic on any network will be only due to the content of that network. As here production stuff is sports and concerts videos, so as much as good such stuff will be there traffic on the website will increase. And as the traffic increases need of server also increases. For the setup of servers a huge capital will be required, but the efficient use of resources available such investments may get reduces up to a great extent. Management of web contents is the one thing and handling the upcoming traffic is the other one. Data can managed by the way discussed above by using some content management architectures like RAID etc. but for handling the traffic high-availability servers are to be maintained. Along with high-availability website mirroring can also be done which may also serve a huge traffic and make the end-user satisfying for the contents. For the high availability of server it will be better if along with data servers if we maintain certain load balancing servers also, which going to handle the expected huge bulk of users coming on website. Here we need to maintain more than one datacentres in parallel with each other and Load balancing server divides the traffic on the website after it reaches the maximum number of requests allotted to get handled by a particular server. When requests goes on increment extra traffic will be diverted to the next datacentre server which responded as same as like the main server as both are of same configuration and with the same content available on both.

For this there is no need of such a huge bulk of supporting staff but instead of this only four to five individuals will be sufficient of the administration and regulation of that server. By such architecture production staff and supporting staff gets reduces and company goes to a huge profit.

Partition of data: In the data partition there will be requirement of email, database and website. As we are going to maintain our own server so such requirements are going to be fulfilled very easily as for website we need a domain and on that domain we can easily allocate an email id for any user or for our staff. For database we are itself having a huge amount of data storage for concert and sports videos so some space form there can be used as our database too. By doing so, we can maintain log and administrate the usage of data by all the end-users and management teams. As we only have to design such an architecture which shows us the content regulation by each user and management staff. This may also be an advance feature for the website which increases the traffic on this.

Storage: We are storing videos of our concerts and sports events of stadium. Videos of the whole event will be quite huge in size and there may be such possibility that number of videos will be there for a single event. So, there should be such a space in size that videos get easily stored without any problem. As per now each video is allocated a disk size of 2GB which is a quite huge amount for a single video. In case of being more than one video total number of videos can be stored at any time should kept as less than 1000. There is an especial privilege for the staff as they will get a disk space of 5GB instead of 2GB. Such, a privilege motivate end-users to be a member staff of the organization which definitely help it one or the other way.

Remote Access and Security:

Data and network security plays a very vital role in web world as, with the increase in reputation the risk factor also gets increases to maintain the integrity. So, firstly the network should be secure enough that there will be no unauthorized access in the network and other one is the security of the data content. Mainly, the network of production staff and admin/management staff both should be 100% secure both on site and also remotely. Because any unauthorized data access may results in a huge loss to the firm. Once if any unauthorized entrance will be there then it may create much more loop whole for future use to get enter in the network. So it is quite necessary to build such a safe and non-penetrating network architecture that no one can get access to it and if tried to do so get easily traced. Tracing of such an activity is also very important because by doing so we may get much more information that what are the flaws available in our existing network and server architectures which help in improving our existing setups. So, firstly our network should be secure enough that no one gets enter into it and next is about our servers, as our servers will be online and will accessible around the globe. Anyone from any part of the world may attack our servers and try to make them compromised. So we must be ready for such an activities, this can be done by keeping continuous log of website and all other type of services provided by our servers. Maintenance of log files for each and every activity on website also helps in improvement of contents and satisfying our clients as per their desire which they want from our servers to serve. We can get to know about the desire of our client by analysing the type of contents searched by them on the site and by the type of content they are usually used to access. So from the next time we may develop such an search engine which works on the basis of logs of each users.

Backup: As day by day our data sets are goes on increasing and due to being online 24x7 we cannot let our servers down even for a single moment. So, it is not a good idea to keep all the data at a single place, all the data should be kept at more than one place and the main data should be backed up time to time which keeps the integrity of content and also maintain. Backup of maim content can be taken on Tape or on Discs on daily, weekly, monthly or yearly basis. But as the backup is taken after short interval like week so, any type of miss happening is

recoverable up to maximum extent as we will have maximum data between two backup spans. Short span backups will be seems to be better in case of any miss happenings if happens anyhow. Because if backups are if small interval then it will be quite easy to recover the maximum of data available and if maximum data is being recovered if will be better beneficial for the firm. Backup of server are done by keeping there kickstart files in which there whole setup configuration are there. By the help of that kickstart configuration file we can easily setup the whole server architecture again in a short while.

Summary: For the organizations like IFPC where there are distributed work architecture, are firstly must design there communication architecture very well because in such organization which provides any sort of service over internet should have high performing network and essentially their website should be high availability factor 24x7. As IFPC is providing service of videos over internet with the recordings of concerts and sports and as per analysis there are a huge bulk of internet search are there for videos available online regarding any concert or of any sports events. Even after a long time also search of any content is not going to be decrease because no one knows when there will be need of any type of data available. So, by keeping such things also we have to be preparing with our data centre and with website any search on website should respond to the best of anyone can. Such a service will definitely increase the popularity of website and also populate it with the huge amount of users. For that our online availability is very much important all the time. Any moment if our servers failed to fulfil the requirement of any client that client not going to return to website again. So, firstly there internal network should be on LAN. It will be better as LAN provides a quite good. Keeping servers on LAN make the network connectivity much more reliable and continuous data transmission with a huge transfer rate. For the high availability of server it will be better if along with data servers if we maintain certain load balancing servers also, which going to handle the expected huge bulk of users coming on website. Here we need to maintain more than one datacentres in parallel with each other and Load balancing server divides the traffic on the website after it reaches the maximum number of requests allotted to get handled by a particular server.

When requests goes on increment extra traffic will be diverted to the next datacentre server which responded as same as like the main server as both are of same configuration and with the same content available on both. Next thing is about backups, with the maintenance of such an excellent architecture by maintaining such a servers, we have to keep backup of the whole architecture so that we can easily set it up again if anything wrong happens like disaster. So, backup of server are done by keeping there kickstart files in which there whole setup configuration are there. By the help of that kickstart configuration file we can easily setup the whole server architecture again in a short while. And in case of datacentre we need to keep backup of the whole system time to time and keeping in mind that this backup time interval should be small. Because small interval works more satisfactorily during data recovery and gives the maximum data to get recovered.

Potrebbero piacerti anche