Modernisation of data architecture is key to maximising value across the business.
Dietmar Rietsch, CEO of Pimcore, identifies best practices for organisations to consider when managing modern enterprise data architecture
Time and again, data has been touted as the lifeline that businesses need to grow and, more importantly, differentiate and lead. Data powers decisions about their business operations and helps solve problems, understand customers, evaluate performance, Improve processes, measure improvement, and much more. However, having data is just a good start. Businesses need to manage this data effectively to put it into the right context and figure out the “what, when, who, where, why and how” of a given situation to achieve a specific set of goals. Evidently, a global, on-demand enterprise survives and thrives on an efficient enterprise data architecture that serves as a source of product and service information to address specific business needs.
A highly functional product and master data architecture is vital to accelerate the time-to-market, Improve customer satisfaction, reduce costs, and acquire greater market share. It goes without saying that data architecture modernisation is the true endgame to meet today’s need for speed, flexibility, and innovation. Now living in a data swamp, enterprises must determine whether their legacy data architecture can handle the vast amount of data accumulated and address the current data processing needs. Upgrading their data architecture to Improve agility, enhance customer experience, and scale fast is the best way forward. In doing so, they must follow best practices that are critical to maximising the benefits of data architecture modernisation.
Below are the seven best practices that must be followed for enterprise data architecture modernisation.
Enterprises gain a potent competitive edge by enhancing their ability to explore data and leverage advanced analytics. To achieve this, they are shifting toward denormalised, mutable data schemas with lesser physical tables for data organisation to maximise performance. Using flexible and extensible data models instead of rigid ones allows for more rapid exploration of structured and unstructured data. It also reduces complexity as data managers do not need to insert abstraction layers, such as additional joins between highly normalised tables, to query relational data.
Data architects are moving away from clusters of centralised enterprise data lakes to domain-based architectures. Herein, data virtualisation techniques are used throughout enterprises to organise and integrate distributed data assets. The domain-driven approach has been instrumental in meeting specific business requirements to speed up the time to market for new data products and services. For each domain, the product owner and product team can maintain a searchable data catalog, along with providing consumers with documentation (definition, API endpoints, schema, and more) and other metadata. As a bounded context, the domain also empowers users with a data roadmap that covers data, integration, storage, and architectural changes.
This approach significantly reduces the time spent on building new data models in the lake, usually from months to days. Instead of creating a centralised data platform, organisations can deploy logical platforms that are managed within various departments across the organisation. For domain-centric architecture, a data infrastructure as a platform approach leverages standardised tools for the maintenance of data assets to speed up implementation.
Implications of data silos for the data-driven enterprise are diverse. Due to data silos, business operations and data analytics initiatives are hindered since it is not possible to interpret unstructured, disorganised data. Organisational silos make it difficult for businesses to manage processes and make decisions with accurate information. Removing silos allows businesses to make more informed decisions and use data more effectively. Evidently, a solid enterprise architecture must eliminate silos by conducting an audit of internal systems, culture, and goals.
A crucial part of modernising data architecture involves making internal data accessible to the people who need it when they need it. When disparate repositories hold the same data, data duplicates created make it nearly impossible to determine which data is relevant. In a modern data architecture, silos are broken down, and information is cleansed and validated to ensure that it is accurate and complete. In essence, enterprises must adopt a complete and centralised MDM and PIM to automate the management of all information across diverse channels in a single place and enable the long-term dismantling of data silos.
With the advent of real-time product recommendations, personalised offers, and multiple customer communication channels, the business world is moving away from legacy systems. For real-time data processing, modernising data architecture is a necessary component of the much-needed digital transformation. With a real-time architecture, enterprises can process and analyse data with zero or near-zero latency. As such, they can perform product analytics to track behaviour in digital products and obtain insights into feature use, UX changes, usage, and abandonment.
The deployment of such an architecture starts with the shift from a traditional model to one that is data-driven. To build a resilient and nimble data architecture model that is both future-proof and agile, data architects must integrate newer and better data technologies. Besides, streaming models, or a combination of batch and stream processing, can be deployed to solve multiple business requirements and witness availability and low latency.
Data today is no longer limited to structured data that can be analysed with traditional tools. As a result of big data and cloud computing, the sheer amount of structured and unstructured data holding vital information for businesses is often difficult to access for various reasons. It implies that the data architecture should be able to handle data from both structured and unstructured sources, both in a structured and an unstructured format. Unless enterprises do so, they miss out on essential information needed to make informed business decisions.
Data can be exposed through APIs so that direct access to view and modify data can be limited and protected, while enabling faster and more current access to standard data sets. Data can be reused among teams easily, accelerating access to and enabling seamless collaboration among analytics teams. By doing this, AI use cases can be developed more efficiently.
Cloud computing is probably the most significant driving force behind a revolutionary new data architecture approach for scaling AI capabilities and tools quickly. The declining costs of cloud computing and the rise of in-memory data tools are allowing enterprises to leverage the most sophisticated advanced analytics. Cloud providers are revolutionising how companies of all sizes source, deploy and run data infrastructure, platforms, and applications at scale. With a cloud-based PIM or MDM, enterprises can take advantage of ready-use and configured solutions, wherein they can seamlessly upload their product data, automate catalog creation, and enrich it diverse marketing campaigns.
With a cloud PIM or MDM, enterprises can eliminate the need for hardware maintenance, application hosting, version updates, and security patches. From the cost perspective, the low cost of subscription of cloud platforms is beneficial for small businesses that can scale their customer base cost-effectively. Besides, cloud-based data platforms also bring a higher level of control over product data and security.
Businesses often have to move beyond legacy data ecosystems offered by prominent solution vendors to scale applications. Many organisations are moving toward modular data architectures that use the best-of-breed and, frequently, open source components that can be swapped for new technologies as needed without affecting the other parts of the architecture. An enterprise using this method can rapidly deliver new, data-heavy digital services to millions of customers and connect to cloud-based applications at scale. Organisations can also set up an independent data layer that includes commercial databases and open source components.
Data is synchronised with the back-end systems through an enterprise service bus, and business logic is handled by microservices that reside in containers. Aside from simplifying integration between disparate tools and platforms, API-based interfaces decrease the risk of introducing new problems into existing applications and speed time to market. They also make the replacement of individual components easier.
Modernising data architecture allows businesses to realise the full value of their unique data assets, create insights faster through AI-based data engineering, and even unlock the value of legacy data. A modern data architecture permits an organisation’s data to become scalable, accessible, manageable, and analysable with the help of cloud-based services. Furthermore, it ensures compliance with data security and privacy guidelines while enabling data access across the enterprise. Using a modern data approach, organisations can deliver better customer experiences, drive top-line growth, reduce costs, and gain a competitive advantage.
How to get ahead of the National Data Strategy to drive business value — Toby Balfre, vice-president, field engineering EMEA at Databricks, discusses how organisations can get ahead of the National Data Strategy to drive business value.
A guide to IT governance, risk and compliance — Information Age presents your complete business guide to IT governance, risk and compliance.
The MarketWatch News Department was not involved in the creation of this content.
Jul 08, 2022 (Heraldkeepers) -- New Jersey, United States-The Datacentre Network Architecture market study contains huge examination information and verifications and is planned to be a valuable asset report for supervisors, investigators, industry specialists, and other key individuals who need a prepared to-get to, self-broke down study to more readily comprehend market patterns, development drivers, potential open doors, and impending difficulties, as well as about contenders.
Receive the trial Report of Datacentre Network Architecture Market 2022 to 2030:
key points of a market Research Report:
• A top to bottom examination of the market on an overall and provincial level is remembered for the market report.
• Critical changes in market elements and rivalry.
• Division by kind, application, geology, and different standards.
• Statistical surveying, both verifiable and future, concerning size, share development, volume, and deals.
• Critical changes in market elements and improvements, as well as appraisals.
• Key portions and districts that are on the ascent
• Significant market members’ center business systems, as well as their key methodologies.
The worldwide Datacentre Network Architecture market is expected to grow at a booming CAGR of 2022-2030, rising from USD billion in 2021 to USD billion in 2030. It also shows the importance of the Datacentre Network Architecture market main players in the sector, including their business overviews, financial summaries, and SWOT assessments.
Datacentre Network Architecture Market Segmentation & Coverage:
Datacentre Network Architecture Market segment by Type:
Datacentre Network Architecture Market segment by Application:
Pharmaceuticals, Life Sciences, Automobile, IT & Telecom, Public, BFSI, Others
The years examined in this study are the following to estimate the Datacentre Network Architecture market size:
History Year: 2015-2019
Base Year: 2021
Estimated Year: 2022
Forecast Year: 2022 to 2030
Cumulative Impact of COVID-19 on Market:
Coronavirus can possibly have three significant monetary results: COVID-19 can possibly have three fundamental outcomes on the worldwide economy: straightforwardly influencing creation and request, upsetting stockpile chains and commercial centers, and influencing undertakings and monetary business sectors monetarily. The COVID-19 pandemic helpfully affects market development, as reception has expanded to all the more likely to understand the monetary effect of COVID-19.
Get a trial Copy of the Datacentre Network Architecture Market Report: https://www.infinitybusinessinsights.com/request_sample.php?id=837340
The Asia-Pacific district has as of late overwhelmed the global Datacentre Network Architecture market, attributable to the inescapable as an industry. During the survey time frame, the APC region is supposed to keep up with its market predominance.
The Key companies profiled in the Datacentre Network Architecture Market:
The study examines the Datacentre Network Architecture market’s competitive landscape and includes data on important suppliers, including Cisco, Juniper Networks, Arista Networks, Hewlett-Packard, Dell, Brocade Communications, IBM, Avaya Networks,& Others
Table of Contents:
List of Data Sources:
Chapter 2. Executive Summary
Chapter 3. Industry Outlook
3.1. Datacentre Network Architecture Global Market segmentation
3.2. Datacentre Network Architecture Global Market size and growth prospects, 2015 – 2026
3.3. Datacentre Network Architecture Global Market Value Chain Analysis
3.3.1. Vendor landscape
3.4. Regulatory Framework
3.5. Market Dynamics
3.5.1. Market Driver Analysis
3.5.2. Market Restraint Analysis
3.6. Porter’s Analysis
3.6.1. Threat of New Entrants
3.6.2. Bargaining Power of Buyers
3.6.3. Bargaining Power of Buyers
3.6.4. Threat of Substitutes
3.6.5. Internal Rivalry
3.7. PESTEL Analysis
Chapter 4. Datacentre Network Architecture Global Market Product Outlook
Chapter 5. Datacentre Network Architecture Global Market Application Outlook
Chapter 6. Datacentre Network Architecture Global Market Geography Outlook
6.1. Datacentre Network Architecture Industry Share, by Geography, 2022 & 2030
6.2. North America
6.2.1. Market 2022 -2030 estimates and forecast, by product
6.2.2. Market 2022 -2030, estimates and forecast, by application
6.2.3. The U.S.
220.127.116.11. Market 2022 -2030 estimates and forecast, by product
18.104.22.168. Market 2022 -2030, estimates and forecast, by application
22.214.171.124. Market 2022 -2030 estimates and forecast, by product
126.96.36.199. Market 2022 -2030, estimates and forecast, by application
6.3.1. Market 2022 -2030 estimates and forecast, by product
6.3.2. Market 2022 -2030, estimates and forecast, by application
188.8.131.52. Market 2022 -2030 estimates and forecast, by product
184.108.40.206. Market 2022 -2030, estimates and forecast, by application
6.3.4. the UK
220.127.116.11. Market 2022 -2030 estimates and forecast, by product
18.104.22.168. Market 2022 -2030, estimates and forecast, by application
22.214.171.124. Market 2022 -2030 estimates and forecast, by product
126.96.36.199. Market 2022 -2030, estimates and forecast, by application
Chapter 7. Competitive Landscape
Chapter 8. Appendix
What are the various types of worldwide business sectors?
Who are the Datacentre Network Architecture market’s critical players?
What locales are impacted by the Datacentre Network Architecture market?
What stages does the worldwide Datacentre Network Architecture market go through?
International: +1 518 300 3575
The MarketWatch News Department was not involved in the creation of this content.
New Jersey, N.J., Aug 03, 2022 The Secure Access Service Edge (SASE) Market Research Report is a professional asset that provides dynamic and statistical insights into regional and global markets. It includes a comprehensive study of the current scenario to safeguard the trends and prospects of the market. Secure Access Service Edge (SASE) Research reports also track future technologies and developments. Thorough information on new products, and regional and market investments is provided in the report.
Secure Access Service Edge (SASE) is a network architecture that combines VPN and SD-WAN features with cloud-native security features such as secure internet gateways, cloud access security brokers, firewalls, and zero trust network access.
Advances in cloud computing technologies are helping to increase business productivity and strengthen security network management. Realizing the benefits of cloud computing, companies are aggressively deploying cloud-based IT infrastructure. Therefore, the growing popularity of cloud-based IT systems and solutions also bodes well for the growth of the market.
Get the PDF trial Copy (Including FULL TOC, Graphs, and Tables) of this report @:
“Secure Access Service Edge (SASE) is growing at a good CAGR over the forecast period. Increasing individual interest in Service industry is a major reason for the expansion of this market.”
Top Companies in this report are:
Cisco Systems, VMware, Fortinet, Inc , Palo Alto Networks, Akamai Technologies, Zscaler,, Cloudflare, Cato Networks (Israel), Versa Networks,, Forcepoint , Broadcom, Check Point Software Technologies Ltd. (Israel), McAfee, LLC , Citrix Systems, Netskope , Perimeter 81 Ltd. (Israel), Open Systems (Switzerland), Aryaka Networks, Proofpoint, Secucloud Network GmbH (Deutschland), Aruba Networks , Juniper Networks, Verizon Communications,, SonicWall , Barracuda Networks, and Twingate ., .
* The report analyses regional growth trends and future opportunities.
* Detailed analysis of each segment provides relevant information.
* The data collected in the report is investigated and Tested by analysts.
* This report provides realistic information on supply, demand and future forecasts.
Secure Access Service Edge (SASE) Market Overview:
This systematic research study provides an inside-out assessment of the Secure Access Service Edge (SASE) market while proposing significant fragments of knowledge, chronic insights and industry-approved and measurably maintained Service market conjectures. Furthermore, a controlled and formal collection of assumptions and strategies was used to construct this in-depth examination.
The report offers an in-depth assessment of the Secure Access Service Edge (SASE) market strategies, and geographic and business segments of the key players in the market.
Market Segmentation: By Type
Network as a service
Security as a service
Market Segmentation: By Application
Retail and eCommerce
IT and ITeS
During the development of this Secure Access Service Edge (SASE) research report, the driving factors of the market are investigated. It also provides information on market constraints to help clients build successful businesses. The report also addresses key opportunities.
For Any Query or Customization: https://a2zmarketresearch.com/ask-for-customization/580912
This report provides an in-depth and broad understanding of Secure Access Service Edge (SASE). With accurate data covering all the key features of the current market, the report offers extensive data from key players. An audit of the state of the market is mentioned as accurate historical data for each segment is available during the forecast period. Driving forces, restraints, and opportunities are provided to help provide an improved picture of this market investment during the forecast period 2022-2029.
Some essential purposes of the Secure Access Service Edge (SASE) market research report:
o Vital Developments: Custom investigation provides the critical improvements of the Secure Access Service Edge (SASE) market, including R&D, new item shipment, coordinated efforts, development rate, partnerships, joint efforts, and local development of rivals working in the market on a global scale and regional.
o Market Characteristics: The report contains Secure Access Service Edge (SASE) market highlights, income, limit, limit utilization rate, value, net, creation rate, generation, utilization, import, trade, supply, demand, cost, part of the industry in general, CAGR and gross margin. Likewise, the market report offers an exhaustive investigation of the elements and their most exact patterns, along with Service market fragments and subsections.
o Investigative Tools: This market report incorporates the accurately considered and evaluated information of the major established players and their extension into the Secure Access Service Edge (SASE) market by methods. Systematic tools and methodologies, for example, Porter’s Five Powers Investigation, Possibilities Study, and numerous other statistical investigation methods have been used to analyze the development of the key players working in the Secure Access Service Edge (SASE) market.
o Convincingly, the Secure Access Service Edge (SASE) report will provide you an unmistakable perspective on every single market reality without the need to allude to some other research report or source of information. This report will provide all of you with the realities about the past, present, and eventual fate of the Service market.
Buy Exclusive Report @: https://www.a2zmarketresearch.com/checkout
1887 WHITNEY MESA DR HENDERSON, NV 89014
+1 775 237 4157
Hybrid and multicloud computing environments have redefined the trust boundary.
In the computer world, a trust boundary serves as an interface for the marking on a data packet that is allowed to flow through a network. Remote work by remote users and the consumption of cloud-based tools to perform business functions have dramatically changed the business environment and the trust boundary along with it.
“The traditional trust boundary has evaporated, or at least transformed dramatically,” said Eric Kostlan (pictured), technical marketing engineer at Cisco Systems Inc. “Although the concept of a trust boundary still exists, the nature of the hybrid, multicloud environment makes it very difficult to define. It’s not that the concept of trusted versus untrusted has gone away; it’s just become fundamentally more complex. The complexity itself is a vulnerability.”
Kostlan spoke with theCUBE industry analysts John Furrier and Dave Vellante at AWS re:Inforce, during an exclusive broadcast on theCUBE, SiliconANGLE Media’s livestreaming studio. They discussed Cisco’s portfolio of security solutions and the need for seamless cloud integration. (* Disclosure below.)
The changing nature of the trust boundary is one of many factors in enterprise computing that Kostlan and his colleagues at Cisco are managing. One of the company’s solutions involves Snort 3, an open-source network security tool for intrusion detection. As more companies have turned to the cloud, tools such as Snort 3 have become key elements that can be integrated in virtual environments.
“There’s a large number of components to the solution, and this spans workload protection, as well as infrastructure protection,” Kostlan said. “These are integrated into cloud components, and this is what allows comprehensive protection across the hybrid cloud environment. Some of the most important technologies that we use, such as Snort 3 — which is a best-of-breed intrusion protection system that we have adopted — are applicable, as well, to the virtual environment so that we push into the cloud in a way that’s seamless.”
Cisco also applies its cloud security solutions by leveraging threat information through its Talos Intelligence Group. Talos is comprised of an experienced group of security experts whose mission is to protect Cisco customer products and services.
“Talos updates our products approximately once every hour with new information about emerging attacks,” Kostlan said. “That architecture is very easily extensible into the cloud, because you can inform a virtual device just as easily as you can inform a physical device of an emergent threat. We have expanded our capacity to visualize what’s happening.”
Here’s the complete video interview, part of SiliconANGLE’s and theCUBE’s coverage of the AWS re:Inforce event:
(* Disclosure: Cisco Systems Inc. sponsored this segment of theCUBE. Neither Cisco nor other sponsors have editorial control over content on theCUBE or SiliconANGLE.)
Concerns are growing among the open source community around big tech's increasing involvement.
Dima Lazerka, co-founder of VictoriaMetrics, discusses how the open source model community can be safeguarded amidst increasing big tech involvement
Free and open source software (FOSS) is an integral part of most of the tech we now use on a daily basis. Originally, it was developed by volunteer developers, however in the last few years there’s been an uptake in the active role of big corporations in the open source model. Deals for Red Hat, Github and Microsoft are the most well-known, but big tech providers across the globe are either incorporating open source into their stacks, or releasing internal technology to the public, such as Spotify open-sourcing Backstage.
However, as big tech becomes increasingly involved in open source, developers have voiced concern about its engagement as it could jeopardise the original ethos and values of the community. While debate is likely to rage on about whether big tech’s involvement is a help or hindrance, open source projects need to be able to maintain freedom of choice.
Big tech has engaged with the open source model mainly by either assigning employees to contribute to existing open source projects, or open-sourcing their own code both to allow the community to utilise it and to help maintain it. Organisations are making open source part of their business model and therefore, have started acquiring an array of open source companies.
Research of 5,800 individuals surveyed by cloud provider DigitalOcean found Google, IBM (including Red Hat) and Microsoft were at the top of a list of 11 companies in terms of open source community citizenship. However, 60 per cent of the surveyed individuals stated they worry about the corporations’ intentions when acquiring or engaging with open source, and 56 per cent mentioned restrictive licences’ role in creating an unfair competitive advantage.
The fact that open source has been made more “corporate” has caused an array of concerns. For starters, acquisitions of open source could potentially result in a crowding out of volunteer developers, jeopardising the future of the open source community. While this isn’t inherently a bad thing as long as big tech’s involvement strengthens the community, open source developers should be able to successfully build their projects without intervention if they so choose.
Open source was founded on a sense of community and collaboration. If developers become disillusioned with these principles, it could have a knock on effect for the rest of the community. Therefore, we must carefully evaluate how to go about building the future of open source communities.
Two of the main techniques to safeguard open source and its community are through smart licensing tactics and constant innovation. The first technique is to simply switch the project licence from an open source licence to a more restrictive licence. There are two specific licences that can be used to protect against clouds and corporations: AGPL-3 and SSPL — specifically developed by the likes of MongoDB, Elastic and Grafana to protect themselves from AWS.
For instance, while many projects shifted away from GPL-style licences towards more permissive forms of licensing, under GPL, contributors are required to make their code available to the open source community; the so-called “copyleft”. This traditional licensing style helps to create a more open, transparent ecosystem.
Another way in which open source can safeguard its future is through smart innovations. Constantly innovating in order to satisfy users should be the way forward for the evolution of open source projects and solutions. This would enable companies to maintain their competitive edge and keep up with technological trends. The beauty of open source is that it is made up of a large ecosystem of innovators in itself, and rather than competing for knowledge, resources are shared for others to benefit from and keep innovating. This has to remain integral to FOSS as it has been the key driver of innovation and growth for FOSS organisations since the beginning.
Ultimately, big tech involvement is not necessarily harmful to the FOSS community, it could actually help it reach its full potential if precautions are taken and the freedom of open source developers is safeguarded and prioritised.
Open source, diversity and inclusion: is the community doing enough? — Ann Schlemmer, President of Percona, asks if the open source community doing a good enough job around diversity and inclusion?
WIT Q&A: digital transformation and open source — Leslie Hawthorn, vertical community strategy manager at Red Hat, and Cali Dolfi, data scientist at Red Hat, spoke to Information Age about digital transformation trends in open source, and promoting workplace DEI.
There’s a problem in the
unrar utility, and as a result, the Zimbra mail server was vulnerable to Remote Code Execution by simply sending an email. So first,
unrar is a source-available command-line application made by RarLab, the same folks behind WinRAR. CVE-2022-30333 is the vulnerability there, and it’s a classic path traversal on archive extraction. One of the ways this attack is normally pulled off is by extracting a symlink to the intended destination, which then points to a location that should be restricted.
unrar has code hardening against this attack, but is sabotaged by its cross-platform support. On a Unix machine, the archive is checked for any symbolic links containing the
../ pattern. After this check is completed, a function runs to convert any Windows paths to Unix notation. As such, the simply bypass is to include symlinks using
..\ traversal, which don’t get caught by the check, and then are converted to working directories.
That was bad enough, but Zimbra made it worse by automatically extracting
.rar attachments on incoming emails, in order to run a virus and spam check. That extraction isn’t sandboxed, so an attacker’s files are written anywhere on the filesystem the
zimbra user can write. It’s not hard to imagine how this turns into a full RCE very quickly. If you have an
unrar binary based on RarLab code, check for version 6.1.7 or 6.12 of their binary release. While Zimbra was the application specifically called out, there are likely to be other cases where this could be used for exploitation.
A widespread malware campaign has been discovered in a bit of an odd place: running on the firmware of Small Office/Home Office (SOHO) network devices. The surprising element is how many different devices are supported by the campaign, including Cisco, Netgear, ASUS, etc. The key is that the malware is a little binary compiled for the MIPS architecture, which is used by many routers and access points.
Once in place, the malware then launches Man-in-the-Middle attacks against DNS and HTTP connections, all with the goal of compromising the Windows machines that run their connection through the compromised device. There have been mass exploit campaigns in the past, where the DNS resolver was modified on vulnerable routers, but this seems to be quite a bit more sophisticated, leading the researchers to suspect that this may be a state-sponsored campaign. There’s an odd note in the source report, that the initial exploit script makes a call to
/cgi-bin/luci, which is the web interface for OpenWRT routers. We’ve reached out for more information from Lumen, so stay tuned for the details. It may very well be that this malware campaign is specifically targeting the hoard of very old, vulnerable OpenWRT-based routers out there. There may be a downside to multiple companies using old versions of the Open Source project as their SDK.
Now it’s a bit different from how a standard overflow exploit works. One reason, Wasm doesn’t have address layout randomization or Data Execution Prevention. On the other hand, web assembly functions don’t reside at a memory address, but simply a function index. The
RET instruction equivalent can’t jump to arbitrary locations, but just to function indexes. However, it’s still a stack, and overflowing a buffer can result in overwriting important data, like the return pointer. Time will tell whether WebAssembly exploits are going to be a big deal, or will forever be a novelty.
In our new, brave future, remote work seems to be the new standard, and this brings some new security considerations. Example: Microsoft’s Intune remote management suite. It’s supposed to be an easy way to deploy, manage, and monitor laptops and desktops remotely. In theory, a robust remote admin suite combined with Bitlocker should make for an effective protection against tampering. Why Bitlocker? While it prevents an attacker from memorizing data from the disk, it also prevents tampering. For instance, there’s a really old trick, where you copy the
cmd.exe binary over the top of the sticky keys, or accessibility binary. These can be launched from the login page, and results in a super-easy root shell. Bitlocker prevents this.
It sounds great, but there’s a problem. Intune can be deployed in two ways. The “user-driven” flow results in a system with more administrative capabilities entrusted to the end user, including access to the BitLocker recovery key. The only way around this is to do the setup, and then remove the Primary User and rotate the Bitlocker keys. Then there’s the troubleshooting mode, holding Shift+F10 during initial setup grants SYSTEM access to the end user. Yikes. And finally, that last gotcha to note is that a remote wipe removes user data, and deletes extra binaries from some important places, but doesn’t do any sort of file verification, so our simple sticky-keys hack would survive. Oof.
[Jack Dates] participated in 2021 Pwn2Own, and put together an Apple Safari exploit that uses the Intel graphics kernel extensions for the escape. It’s a *very* deep dive into OSX exploitation. The foothold is an off-by-one error in a length check, which in total allows writing four bytes of arbitrary data. The key to turn this into something useful was to strew some corpses around memory — forked, dead processes. Corrupt the size of the corpse, and you can use it to free other memory that’s still in use. Use after free for the win!
The OpenSSL bug we talked about last week is still being looked into, with [Guido Vranken] leading the charge. He found a separate bug that specifically isn’t a security problem back in May, and it’s the fix for that bug that introduced the AVX512 problem we’re interested in. There still looks to be a potential for RCE here, but at least it’s proving to be non-trivial to put such an attack together.
There’s a new malware campaign, ytstealer, that is explicitly targeting YouTube account credentials. The malware is distributed as fake installers for popular tools, like OBS Studio, Auto-Tune, and even cheats and cracks for other software. When run, YTStealer looks for an authentication cookie, logs into YouTube Studio, and grabs all the data available there from the attached account. This information and cookie are encrypted and sent on to a C&C server. It’s unclear why YouTube accounts are so interesting to an attacker, but maybe we can all look forward to spammy videos getting uploaded to our favorite channels.
And finally, because there’s more to security than just computers, a delightful puzzle solve from LockPickingLawyer. Loki is a puzzle lock made from a a real padlock, and it caught the attention of our favorite lock-picker, who makes an attempt to open it. We won’t spoil any of the results, but if puzzles or locks are your jam, it’s worth a watch.
SAN FRANCISCO--(BUSINESS WIRE)--Aug 3, 2022--
Fastly, Inc. (NYSE: FSLY), the world’s fastest global edge cloud platform, today announced that the Board of Directors has appointed Todd Nightingale as the company’s next Chief Executive Officer, effective September 1, 2022. Nightingale will also join the Fastly Board of Directors upon assuming the role. He will succeed Joshua Bixby, who, as previously announced, will step down as CEO and from Fastly’s Board of Directors. Bixby will remain with Fastly as an advisor.
This press release features multimedia. View the full release here: https://www.businesswire.com/news/home/20220803005944/en/
Fastly Appoints Todd Nightingale as CEO (Photo: Business Wire)
Nightingale’s appointment culminates a broad search process to identify the company’s next leader. He joins Fastly from Cisco, where he currently leads business strategy and development efforts for Cisco's multi-billion dollar networking portfolio as Executive Vice President and General Manager of Enterprise Networking and Cloud.
“Todd is a proven and passionate technology leader and we are thrilled to have him join our team,” said David Hornik, Lead Independent Director on the Fastly Board of Directors. “We are confident that Todd’s deep background helping customers transform their infrastructures and digitize their businesses will be instrumental to strengthening Fastly’s technology and go-to-market strategy and lead the company into its next stage of growth.”
"Fastly is extraordinary at the things that make us unique, including our incredibly powerful programmable edge cloud, innovative performance-focused product and engineering, and our unmatched support of customers as they build the next generation of globally performant, secure and reliable applications," said Artur Bergman, Fastly’s Founder, Chief Architect and Executive Chairperson. "I'm confident in Todd's ability to lead the company with the rigor and energy needed to elevate Fastly to its next level of extraordinary technology and product growth, including a strong go-to-market motion and operational strengths."
“Fastly is delivering unparalleled application experiences for users around the world with exceptional flexibility, security and performance,” said Nightingale. “I'm honored and grateful for the opportunity to be a part of the Fastly team.”
During his time at Cisco, Todd Nightingale led the Enterprise Networking and Cloud business as Executive Vice President and General Manager. He managed business strategy and development efforts for Cisco's multi-billion-dollar networking portfolio. Nightingale is known for his passionate technology leadership and his vision of powerful, simple solutions for businesses, schools, and governments. Previously, Nightingale was the Senior Vice President and General Manager of Cisco's Meraki business. His focus on delivering a simple, secure, digital workplace led to the expansion and growth of the Meraki portfolio, making it the largest cloud-managed networking platform in the world. Nightingale joined Cisco with the Meraki acquisition in 2012. He previously held engineering and senior management positions at AirDefense, where he was responsible for product development and guided the company through a successful acquisition by Motorola.
Fastly’s powerful and programmable edge cloud platform helps the world’s top brands deliver the fastest online experiences possible, while improving site performance, enhancing security, and empowering innovation at global scale. With world-class support that consistently achieves 95%+ customer satisfaction ratings*, Fastly's beloved suite of edge compute, delivery, and security offerings has been recognized as a leader by industry analysts such as IDC, Forrester and Gartner. Compared to legacy providers, Fastly’s powerful and modern network architecture is the fastest on the planet, empowering developers to deliver secure websites and apps at global scale with rapid time-to-market and industry-leading cost savings. Thousands of the world’s most prominent organizations trust Fastly to help them upgrade the internet experience, including Reddit, Pinterest, Stripe, Neiman Marcus, The New York Times, Epic Games, and GitHub. Learn more about Fastly at https://www.fastly.com/, and follow us @fastly.
*As of June 1, 2022
This press release contains “forward-looking” statements that are based on Fastly’s beliefs and assumptions and on information currently available to Fastly on the date of this press release. Forward-looking statements may involve known and unknown risks, uncertainties, and other factors that may cause its actual results, performance, or achievements to be materially different from those expressed or implied by the forward-looking statements. These statements include, but are not limited to, those regarding Mr. Nightingale’s anticipated appointment as Chief Executive Officer and a member of Fastly’s Board of Directors, Fastly’s ability to strengthen its technology and go-to-market strategy, enter its next stage of growth, and deliver a robust portfolio for customers to continue developing the next generation of globally performant, secure and reliable applications. Except as required by law, Fastly assumes no obligation to update these forward-looking statements publicly, or to update the reasons actual results could differ materially from those anticipated in the forward-looking statements, even if new information becomes available in the future. Important factors that could cause Fastly’s actual results to differ materially are detailed from time to time in the reports Fastly files with the Securities and Exchange Commission (SEC), in its Annual Report on Form 10-K for the fiscal year ended December 31, 2021. Additional information will also be set forth in Fastly’s Quarterly Report on Form 10-Q for the fiscal quarter ended June 30, 2022. Copies of reports filed with the SEC are posted on Fastly’s website and are available from Fastly without charge.
Source: Fastly, Inc.
View source version on businesswire.com:https://www.businesswire.com/news/home/20220803005944/en/
CONTACT: Investor Contact:
Vernon Essi, Jr.
KEYWORD: CALIFORNIA UNITED STATES NORTH AMERICA
INDUSTRY KEYWORD: NETWORKS INTERNET SECURITY TECHNOLOGY SOFTWARE
SOURCE: Fastly, Inc.
Copyright Business Wire 2022.
PUB: 08/03/2022 04:05 PM/DISC: 08/03/2022 04:07 PM
Were you unable to attend Transform 2022? Check out all of the summit sessions in our on-demand library now! Watch here.
As the world becomes increasingly data-driven, businesses must find suitable solutions to help them achieve their desired outcomes. Data lake storage has garnered the attention of many organizations that need to store large amounts of unstructured, raw information until it can be used in analytics applications.
The data lake solution market is expected to grow rapidly in the coming years and is driven by vendors that offer cost-effective, scalable solutions for their customers.
Learn more about data lake solutions, what key features they should have and some of the top vendors to consider this year.
A data lake is defined as a single, centralized repository that can store massive amounts of unstructured and semi-structured information in its native, raw form.
It’s common for an organization to store unstructured data in a data lake if it hasn’t decided how that information will be used. Some examples of unstructured data include images, documents, videos and audio. These data types are useful in today’s advanced machine learning (ML) and advanced analytics applications.
Data lakes differ from data warehouses, which store structured, filtered information for specific purposes in files or folders. Data lakes were created in response to some of the limitations of data warehouses. For example, data warehouses are expensive and proprietary, cannot handle certain business use cases an organization must address, and may lead to unwanted information homogeneity.
On-premise data lake solutions were commonly used before the widespread adoption of the cloud. Now, it’s understood that some of the best hosts for data lakes are cloud-based platforms on the edge because of their inherent scalability and considerably modular services.
A 2019 report from the Government Accountability Office (GAO) highlights several business benefits of using the cloud, including better customer service and the acquisition of cost-effective options for IT management services.
Cloud data lakes and on-premise data lakes have pros and cons. Businesses should consider cost, scale and available technical resources to decide which type is best.
Read more about data lakes: What is a data lake? Definition, benefits, architecture and best practices
It’s critical to understand what features a data lake offers. Most solutions come with the same core components, but each vendor may have specific offerings or unique selling points (USPs) that could influence a business’s decision.
Below are five key features every data lake should have:
Data lakes that offer diverse interfaces, APIs and endpoints can make it much easier to upload, access and move information. These capabilities are important for a data lake because it allows unstructured data for a wide range of use cases, depending on a business’s desired outcome.
ML engineers, data scientists, decision-makers and analysts benefit most from a centralized data lake solution that stores information for easy access and availability. This characteristic can help data professionals and IT managers work with data more seamlessly and efficiently, thus improving productivity and helping companies reach their goals.
Imagine a data lake with large amounts of information but no sense of organization. A viable data lake solution must incorporate generic organizational methods and search capabilities, which provide the most value for its users. Other features might include key-value storage, tagging, metadata, or tools to classify and collect subsets of information.
Security and access control are two must-have features with any digital tool. The current cybersecurity landscape is expanding, making it easier for threat actors to exploit a company’s data and cause irreparable damage. Only certain users should have access to a data lake, and the solution must have strong security to protect sensitive information.
More organizations are growing larger and operating at a much faster rate. Data lake solutions must be flexible and scalable to meet the ever-changing needs of modern businesses working with information.
Also read: Unlocking analytics with data lake and graph analysis
Some data lake solutions are best suited for businesses in certain industries. In contrast, others may work well for a company of a particular size or with a specific number of employees or customers. This can make choosing a potential data lake solution vendor challenging.
Companies considering investing in a data lake solution this year should check out some of the vendors below.
The AWS Cloud provides many essential tools and services that allow companies to build a data lake that meets their needs. The AWS data lake solution is widely used, cost-effective and user-friendly. It leverages the security, durability, flexibility and scalability that Amazon S3 object storage offers to its users.
The data lake also features Amazon DynamoDB to handle and manage metadata. AWS data lake offers an intuitive, web-based console user interface (UI) to manage the data lake easily. It also forms data lake policies, removes or adds data packages, creates manifests of datasets for analytics purposes, and features search data packages.
Cloudera is another top data lake vendor that will create and maintain safe, secure storage for all data types. Some of Cloudera SDX’s Data Lake Service capabilities include:
Other benefits of Cloudera’s data lake include product support, downloads, community and documentation. GSK and Toyota leveraged Cloudera’s data lake to garner critical business intelligence (BI) insights and manage data analytics processes.
Databricks is another viable vendor, and it also offers a handful of data lake alternatives. The Databricks Lakehouse Platform combines the best elements of data lakes and warehouses to provide reliability, governance, security and performance.
Databricks’ platform helps break down silos that normally separate and complicate data, which frustrates data scientists, ML engineers and other IT professionals. Aside from the platform, Databricks also offers its Delta Lake solution, an open-format storage layer that can Improve data lake management processes.
Domo is a cloud-based software company that can provide big data solutions to all companies. Users have the freedom to choose a cloud architecture that works for their business. Domo is an open platform that can augment existing data lakes, whether it’s in the cloud or on-premise. Users can use combined cloud options, including:
Domo offers advanced security features, such as BYOK (bring your own key) encryption, control data access and governance capabilities. Well-known corporations such as Nestle, DHL, Cisco and Comcast leverage the Domo Cloud to better manage their needs.
Google is another big tech player offering customers data lake solutions. Companies can use Google Cloud’s data lake to analyze any data securely and cost-effectively. It can handle large volumes of information and IT professionals’ various processing tasks. Companies that don’t want to rebuild their on-premise data lakes in the cloud can easily lift and shift their information to Google Cloud.
Some key features of Google’s data lakes include Apache Spark and Hadoop migration, which are fully managed services, integrated data science and analytics, and cost management tools. Major companies like Twitter, Vodafone, Pandora and Metro have benefited from Google Cloud’s data lakes.
Hewlett Packard Enterprise (HPE) is another data lake solution vendor that can help businesses harness the power of their big data. HPE’s solution is called GreenLake — it offers organizations a truly scalable, cloud-based solution that simplifies their Hadoop experiences.
HPE GreenLake is an end-to-end solution that includes software, hardware and HPE Pointnext Services. These services can help businesses overcome IT challenges and spend more time on meaningful tasks.
Business technology leader IBM also offers data lake solutions for companies. IBM is well-known for its cloud computing and data analytics solutions. It’s a great choice if an operation is looking for a suitable data lake solution. IBM’s cloud-based approach operates on three key principles: embedded governance, automated integration and virtualization.
These are some data lake solutions from IBM:
With so many data lakes available, there’s surely one to fit a company’s unique needs. Financial services, healthcare and communications businesses often use IBM data lakes for various purposes.
Microsoft offers its Azure Data Lake solution, which features easy storage methods, processing, and analytics using various languages and platforms. Azure Data Lake also works with a company’s existing IT investments and infrastructure to make IT management seamless.
The Azure Data Lake solution is affordable, comprehensive, secure and supported by Microsoft. Companies benefit from 24/7 support and expertise to help them overcome any big data challenges they may face. Microsoft is a leader in business analytics and tech solutions, making it a popular choice for many organizations.
Companies can use Oracle’s Big Data Service to build data lakes to manage the influx of information needed to power their business decisions. The Big Data Service is automated and will provide users with an affordable and comprehensive Hadoop data lake platform based on Cloudera Enterprise.
This solution can be used as a data lake or an ML platform. Another important feature of Oracle is it is one of the best open-source data lakes available. It also comes with Oracle-based tools to add even more value. Oracle’s Big Data Service is scalable, flexible, secure and will meet data storage requirements at a low cost.
Snowflake’s data lake solution is secure, reliable and accessible and helps businesses break down silos to Improve their strategies. The top features of Snowflake’s data lake include a central platform for all information, fast querying and secure collaboration.
Siemens and Devon Energy are two companies that provide testimonials regarding Snowflake’s data lake solutions and offer positive feedback. Another benefit of Snowflake is its extensive partner ecosystem, including AWS, Microsoft Azure, Accenture, Deloitte and Google Cloud.
Companies that spend extra time researching which vendors will offer the best enterprise data lake solutions for them can manage their information better. Rather than choose any vendor, it’s best to consider all options available and determine which solutions will meet the specific needs of an organization.
Every business uses information, some more than others. However, the world is becoming highly data-driven — therefore, leveraging the right data solutions will only grow more important in the coming years. This list will help companies decide which data lake solution vendor is right for their operations.
Read next: Get the most value from your data with data lakehouse architecture
VentureBeat's mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact. Learn more about membership.
Latest Study on Industrial Growth of Worldwide 5G-Network Infrastructure Market 2022-2027. A detailed study accumulated to offer Latest insights about acute features of the Worldwide 5G-Network Infrastructure Market. The report contains different market predictions related to revenue size, production, CAGR, Consumption, gross margin, price, and other substantial factors. While emphasizing the key driving and restraining forces for this market, the report also offers a complete study of the future trends and developments of the market. It also examines the role of the leading market players involved in the industry including their corporate overview, financial summary and SWOT analysis.
Get Free Exclusive PDF trial Copy of This Research @ https://www.advancemarketanalytics.com/sample-report/199527-global-5g-network-infrastructure-market#utm_source=DigitalJournalShraddha
Some of the key players profiled in the study are:
Huawei (China), Ericsson (Sweden), Samsung (South Korea), Nokia Networks (Finland), ZTE (China), NEC (Japan), CISCO (United States), CommScope (United States), Comba Telecom Systems (Hong Kong) and Alpha Networks (Taiwan).
Scope of the Report of 5G-Network Infrastructure
5G network infrastructure is divided into two types: standalone 5G infrastructures, which have their own cloud-native network core that connects to 5G New Radio (NR) technology, and non-standalone (NSA) infrastructures, which still rely on existing 4G LTE infrastructure to some extent. To provide a 5G-like experience until network carriers can build out the independent infrastructure required for 5G, the NSA approach uses a combination of 5G Radio Access Network (RAN), 5G NR interface, and existing LTE infrastructure and core network. A standalone 5G deployment consists of user equipment, such as the RAN and NR interface, as well as the 5G core network, which is built on a service-based architecture framework with virtualized network functions. Network functions that were previously executed on hardware are now virtualized and executed as software.
The titled segments and sub-section of the market are illuminated below:
by Type (Standalone, Non-Standalone), Application (Critical communications, Enterprise networking, Industrial Internet of Things (IoT), Others), Technology (Software-Defined Networking (SDN), Network Function Virtualization (NFV)), Frequency Band (5G high-band (mmWave), 5G mid-band, 5G low-band), Hardware (Small cells, RAN cell towers, Others), End User (Residential, Commercial, Industrial, Government) Players and Region – Global Market Outlook to 2027
Rapid developments in IoT technology have been contributing to market growth.
Rise number of mobile and internet devices and rapid growth in internet dependency is expected to bring more opportunities for the 5G Infrastructure Market
Have Any Questions Regarding Global 5G-Network Infrastructure Market Report, Ask Our [email protected] https://www.advancemarketanalytics.com/enquiry-before-buy/199527-global-5g-network-infrastructure-market#utm_source=DigitalJournalShraddha
Region Included are: North America, Europe, Asia Pacific, Oceania, South America, Middle East & Africa
Country Level Break-Up: United States, Canada, Mexico, Brazil, Argentina, Colombia, Chile, South Africa, Nigeria, Tunisia, Morocco, Germany, United Kingdom (UK), the Netherlands, Spain, Italy, Belgium, Austria, Turkey, Russia, France, Poland, Israel, United Arab Emirates, Qatar, Saudi Arabia, China, Japan, Taiwan, South Korea, Singapore, India, Australia and New Zealand etc.
Strategic Points Covered in Table of Content of Global 5G-Network Infrastructure Market:
Chapter 1: Introduction, market driving force product Objective of Study and Research Scope the 5G-Network Infrastructure market
Chapter 2: Exclusive Summary – the basic information of the 5G-Network Infrastructure Market.
Chapter 3: Displaying the Market Dynamics- Drivers, Trends and Challenges & Opportunities of the 5G-Network Infrastructure
Chapter 4: Presenting the 5G-Network Infrastructure Market Factor Analysis, Porters Five Forces, Supply/Value Chain, PESTEL analysis, Market Entropy, Patent/Trademark Analysis.
Chapter 5: Displaying the by Type, End User and Region/Country 2015-2020
Chapter 6: Evaluating the leading manufacturers of the 5G-Network Infrastructure market which consists of its Competitive Landscape, Peer Group Analysis, BCG Matrix & Company Profile
Chapter 7: To evaluate the market by segments, by countries and by Manufacturers/Company with revenue share and sales by key countries in these various regions (2021-2027)
Chapter 8 & 9: Displaying the Appendix, Methodology and Data Source
finally, 5G-Network Infrastructure Market is a valuable source of guidance for individuals and companies.
Read Detailed Index of full Research Study at @ https://www.advancemarketanalytics.com/reports/199527-global-5g-network-infrastructure-market#utm_source=DigitalJournalShraddha
Thanks for memorizing this article; you can also get individual chapter wise section or region wise report version like North America, Middle East, Africa, Europe or LATAM, Southeast Asia.
Craig Francis (PR & Marketing Manager)
AMA Research & Media LLP
Unit No. 429, Parsonage Road Edison, NJ
New Jersey USA – 08837
Phone: +1 (206) 317 1218
Cisco and Qwilt today announced the deployment of their unique content delivery solution across the North American network of Kinetic by Windstream, to enable superior streaming performance to its customers throughout the United States. The adoption of Cisco and Qwilt’s open caching solution considerably improves the quality and efficiency of live and on-demand video delivery while increasing Kinetic’s network capacity for other forms of media.
Kinetic will deploy Qwilt’s open caching technology across its 170,000 miles of fiber network across the US. The deployment helps Kinetic address the growing number of live streaming events and other media consumption, as households increasingly turn to streaming services and expect broadcast quality to the home. Cisco’s edge compute and networking infrastructure, combined with Qwilt’s Open Edge Cloud for Content Delivery Solution, makes this possible by preparing Kinetic’s network to support increasing data volumes and improving the streaming experience.
Gary Cooke, Senior Vice President of Engineering for Kinetic, said: “We tell our customers ‘High Speed for Here’ – whether live, on-demand, or other forms of media content. By partnering with Qwilt and Cisco, we’re ensuring our fiber-backed network has the scalability and capacity needed to handle the growth in demand for all forms of streaming content. By integrating innovative edge technologies across our network, we’re bringing high quality content closer to our customers than ever before.”
Open caching, an open architecture developed and endorsed by the Streaming Video Alliance, offers a platform that federates content delivery infrastructure deployed deep inside service provider networks. It provides open APIs, protection, and security mechanisms for content publishers. The open caching approach helps service providers quickly deploy an edge delivery footprint and addresses the needs of global and regional content providers for more capacity, consistency in content delivery, and performance assurance. The deployment also creates a telco cloud foundation for future use cases, such as website delivery and edge computing.
Theodore Tzevelekis, Vice President and Head of Business Development, Mass-scale Infrastructure Group, Cisco, said: “Streaming is the future of content delivery, but it doesn’t have to mark the end of great quality content experiences. By embracing content delivery at the edge through open caching technology, service providers are embracing a new model to manage their network capacity. Working alongside Qwilt, we’re equipping Kinetic by Windstream with the tools needed to bring fantastic experiences to its customers across North America and in doing so, democratize capacity across its nationwide infrastructure.”
Alon Maor, CEO and Co-Founder of Qwilt, said: “Soaring demand for streamed content brings an urgent need for service providers to scale their networks for the future. Through Cisco and Qwilt’s united vision, we empower service providers like Kinetic by Windstream to do more with their assets while accelerating their digital transformations. We look forward to helping modernize the way Kinetic delivers content and ready their growing network for the future of content experiences.”
Qwilt’s mission is to deliver connected experiences at the quality they were imagined. Its model is built on partnerships with service providers and content publishers, globally, to create a fabric that powers high-performance delivery of media and applications at the very edge of neighborhoods, big and small.
Qwilt’s open architecture and inclusive business model make local edge delivery more accessible than ever before, unlocking more reliable, higher quality-of-experience at greater scale than previously possible. A growing number of the world’s leading content publishers and cable, telco, and mobile service providers rely on Qwilt for Edge Cloud services, including BT, Telecom Argentina, Telecom Italia, and Verizon.
Founded in 2010, Qwilt is a leader of the Open Caching movement and a founding member of the Streaming Video Alliance. Qwilt is backed by Accel Partners, Bessemer Venture Partners, Cisco Ventures, Disruptive, Innovation Endeavors, Marker, and Redpoint Ventures. For more information, visit www.qwilt.com.
Cisco (NASDAQ: CSCO) is the worldwide leader in technology that powers the Internet. Cisco inspires new possibilities by reimagining your applications, securing your data, transforming your infrastructure, and empowering your teams for a global and inclusive future. Discover more on The Newsroom and follow us on Twitter.
Cisco and the Cisco logo are trademarks or registered trademarks of Cisco and/or its affiliates in the U.S. and other countries. A listing of Cisco's trademarks can be found at www.cisco.com/go/trademarks. Third-party trademarks mentioned are the property of their respective owners. The use of the word partner does not imply a partnership relationship between Cisco and any other company.
Kinetic by Windstream is a business unit of Windstream Holdings, a privately held FORTUNE 1000 communications and software company. Kinetic provides premium broadband, entertainment and security services through an enhanced fiber network to consumers and businesses primarily in rural areas in 18 states. The company also offers managed communications services, including SD-WAN and UCaaS, and high-capacity bandwidth and transport services to businesses across the U.S. Additional information is available at GoKinetic.com. Follow us on Twitter at @GoKineticHome.
From Fortune. ©2021 Fortune Media IP Limited. All rights reserved. Used under license. Fortune and Fortune 1000 are registered trademarks of Fortune Media IP Limited and are used under license. Fortune is not affiliated with, and does not endorse products or services of, Windstream.