Click an article title to share with others in social media or leave us a comment. If you have any article suggestions you would like us to cover in the future please get in touch and drop us a line.

26Apr 2019

The market for Cloud Services has grown tremendously in recent years; many enterprises started thinking ‘when to migrate their SAP landscape into Cloud Services.’ However, the main topic of discussion among the stakeholders is ‘how to migrate current complex SAP landscape into Cloud Services.’

Let us see how Cloud Comrade expertise can help its customers to migrate SAP Workloads into AWS

Before that let’s have a close look at the benefits of Migrating the Workloads into Cloud over Traditional Workload Migration

Traditional Migration Processes- Manual

  • Assessing the Infrastructure
  • Design
  • Cost Estimation
  • Proof of Concept
  • Migration
  • Testing
  • Operations

Conventionally, it’s challenging for any managed service providers to get 100% visibility of workloads in the infrastructure. Many technologies are outdated and unsupported by phones or tablets which results in a lack of future technical support. The old interface or legacy software could pose a significant security threat, downtime that will affect business productivity and efficiency.

Hence it’s a viable solution for organizations to consider migrating workloads (SAP) into cloud AWS. Like we all know the five R’s involved in the migration to cloud – Rehost, Refactor, Revise, Rebuild or Replace, we can choose based on the necessity.

Migrating SAP Workloads into AWS – Automated:

AWS has been an SAP Global Technology partner since 2011. AWS and SAP have partnered closely in coming up with various services to make fast and efficient deployment of SAP Workloads on AWS for lower costs, innovation, and simplified infrastructure management. With AWS Server Migration Service (SMS), we can migrate on-premises SAP workloads into AWS Cloud quickly and efficiently.

One of the critical benefits migrating SAP Workloads into AWS is the scalability, without worrying about how to accommodate the business growth. The ease of spin up/down/change the configurations.

After the workloads assessment, the service provider should know if there is a need for private/public or hybrid cloud migration. For example, there could be existing legacy systems at its performance limit that limits the SAP upgrades, then its advisable to have a private cloud migration which will help the clients to avoid additional hardware investment and achieve scalability.

Planning and choosing the right service partner is critical. Its critical for the service provider to have a proper understanding of migrating strategy as most of the decision makers struggle in understanding which applications to rehost, as some might require refactoring as well.   

 

Cloud Comrade has a team of professionals who assess existing workloads and provides its customers with the step by step right strategy to migrate SAP Workloads into AWS. We offer a customized cloud infrastructure for highly regulated businesses to maximize their productivity and efficiency

 

Case Study:

Cloud Comrade successfully helped one of their key customers (Security Industry) in migrating their SAP workload from SAP ERP 6.0 on DB2 to SAP ERP 6.0 Suite on HANA using Software Update Manager(SUM) Database Migration Option(DMO) with System Move and SAP HANA Quick Start.

 

Migration Strategy:

26Apr 2019

AWS is the market leader in Cloud Computing. It offers high availability, reliability, efficiency, scalability, and cost-effective solutions to enterprises resulting in high productivity. Hence, we see an increase in the number of enterprises with business-critical SAP landscape systems acclimatizing private/hybrid/ public cloud solutions.

As announced, SAP will extend maintenance support for their customers until the end of 2025 for SAP Business Suite 7 core application releases including SAP ERP 6.0, SAP Customer Relationship Management 7.0, SAP Supply Chain Management 7.0, SAP Supplier Relationship Management 7.0, and SAP Business Suite powered by SAP HANA 2013.

SAP is reforming their traditional support into Next-Generation Support. SAP Next Generation Support is designed for speed incident resolution, helps prevent incidents with proactive resources before they happen.

If the business decides to stick with SAP, they can fully realize all the benefits of SAP S/4HANA and SAP HANA in the AWS Platform

Difference between SAP S/4 HANA and SAP HANA

SAP S/4 HANA stands for SAP Business Suite 4 SAP HANA. It is the next generation business suite built by SAP replacing both the ERP and BI.

In the three Tier architecture of SAP system, SAP HANA is the database layer, SAP S/4 HANA is the application layer, SAPUI5Fiori is the presentation layer.

What is SAP HANA?

SAP HANA Database is a full-fledged, in-memory, relational database which is designed to replicate and ingest structured data from SAP and non-SAP relational databases, applications, and other systems quickly.

SAP HANA is a platform for real-time analysis of big data. It stores all the data in the server memory and processes a large amount of data faster than a traditional database system.

Why opt for SAP HANA in the AWS cloud?

  1. To install SAP HANA in AWS platform is made easy and swift with AWS SAP HANA Quick Start tool (With the Quick Start tool it take just hours compared on traditional on-premises servers installation, where it will take weeks).
  2. AWS has the broadest range of SAP certified instances. It comes with support up to 12TB memory on a single instance and 48TB on four scale-out instances compared to any other public cloud provider.
  3. With the AWS available services, the customers can build secure, highly available, scalable and cost-efficient SAP HANA instances.

 

AWS Well-Architected framework provides the customers to evaluate the architecture and build secure, high-performing, reliable,  cost-effective, resilient, and efficient infrastructure for their SAP applications

The businesses are dealing with the rise of new customized technologies. The systems to be managed range from on-premise to the cloud (public and private), the data is both within the enterprise and outside of it.

The AWS Well-Architected Framework uses a structured approach including strategies to compare the SAP workload with the AWS best practices, to produce secure, stable and efficient systems.

AWS Well Architected Framework focuses on five pillars,

Most of all, choosing the right managed service partner is the key. Cloud Comrade, with its expertise in SAP landscape and HANA, can come up with a customized cloud strategy for high performance, availability, and security requirements for mission-critical systems.

25Feb 2019

It was those old days where were no complex applications, and the servers used to run in the most okay conditions. In this current world, as the technology grows along with the complexity, and the expectation from customers are growing as well. They are looking for one stop automated solution for business as usual.

After the break/fix model became impracticable for business-as-usual, we saw the rise of Managed Services Providers. The objective of the MSP is to increase the productivity of organizations with minimal operating costs.

Then, it was a huge success. All the customers were looking for MSP whom they can rely upon for their entire infrastructure solutions. As days pass by, the complexity of the applications and infrastructure are continuously changing,  and the expectations from the customers as well.

Let’s have a look at Traditional MSP – Manual

  • Monitoring, Management, and Security
  • Storage, Warehouse Management
  • Backup and recovery
  • Authentication
  • Streamlining the Systems/ Applications
  • Reports

Traditional MSP was a manual process, and also the primary concern for clients. The business wanted to have automated systems that can automatically scale up/down, and balance the load with business as usual. As there was an increasing number of customers looking for transparent and automated business technology services, DevOps focused AWS NextGen MSP’s offers enterprises build and deliver applications on AWS.

AWS powered DevOps offers an efficient workflow, that helps businesses automate day to day activities and provide full-lifecycle services to run, and support customers applications and infrastructure. The AWS powered advanced monitoring services which have predictive analysis plus the continuous monitoring, and anomaly detection helps the business to analyze the current facts and predict the future events that the organizations can take proactive measures to increase their productivity and efficiency.  The AWS management reporting helps the business to make data-driven decisions for business performance.

AWS powered DevOps brings remote individuals and in-house teams together and ensures that they are in sync with each other. DevOps focused AWS NextGen MSP help enterprises find the ideal solutions and infrastructure that is cost-effective. It provides efficient business outcomes by continuously assessing and monitoring systems to optimize performance.

Let’s have a look at NextGen MSP – Automated

  • Cloud delivery
  • Security and daily operations
  • 24/7X 365 days IT support
  • Backup and recovery
  • Access management
  • High security
  • Compliance
  • Advanced continuous monitoring and report management

As a Managed Service Provider (MSP) in the AWS environment, Cloud Comrade is committed to building a sustainable automated business powered by AWS, that will continue to support and grow with our clients on the Cloud.

25Feb 2019

Operating a business across multiple environments is challenging enough. Organizations transforming in to digitalization are often surprised by the high costs of their Traditional MSP services and struggle to manage frameworks and governance across the organization. Does any of the above situations apply to your business?

We have an answer; Cloud Comrade is passionate and driven by automated DevOps NextGen AWS Managed Services to ensure Companies productivity gained are by automating their security, infrastructure, software development, and rollout to achieve critical mass in a short turn around time.

For example, let’s talk about the Traditional MSP who does lift and shit, installation, maintenance, network security monitoring, remote and onsite support. Some of the processes are reactive, and in the long run, it’s not sustainable for business operations. Some of the organizations lack technical expertise, so they have to rely more on the traditional MSP’s which wouldn’t be a cost-effective plus proactive method of running the business.

As the complexity of systems increases, the customers are looking for more than just an MSP. The customers are looking for strategic partners who instead of using traditional tools and processes, can automate and proactively provide various solutions for higher-performance computing. With AWS powered DevOps methodologies, we can develop the AWS Security Framework to maintain security and data protection in the cloud. The framework includes security strategy, risk, compliance, governance, security assessments, incident response, and automate threat hunting.

The security analytics and reporting presented by the AWS management reporting is for proactively prioritize and take measures to handle the threat.

The new breed of AWS NextGen Managed Service Provider’s is redefining the Traditional MSP business model and the go-to-market strategies. When compared to the Traditional framework, the AWS NextGen DevOps Transformation framework assesses the organization’s current capability and provides a structured approach to a DevOps transition. With AWS Powered NextGen MSP, the organizations can install Amazon Lex – Build Conversation Bots as a communication medium through B2B/ B2C portals. This build conversation bots are adaptable and can be customized based on the requests from the customers.

25Feb 2019

For the organizations to stay competitive in today’s technology world, have to think of ways to keep their infrastructure automated, highly available, flexible, reproducible, scalable for high productivity and reliability.

In this blog, we will see how the AWS powered NextGen Infrastructure as a code(IaC) helps us to achieve organizations objective compared to the traditional infrastructure as code (IaC).

Traditional Infrastructure as code:

The traditional IaC is to enable and manage the data centers, storage, networking manually. The respective admin will set up the disk, install operating systems and applications. The period required before the launch could be days or weeks. Not only it is time-consuming, but it consumes a big chunk of the workforce plus the higher cost. Imagine of hardware failure, the time required to wait for the manufacturer production, ship, and delivery. What if the hardware malfunctions after all the waiting period. Again the business had to wait for the subject matter expert to handle the situation.

NextGen Infrastructure as code:

With AWS powered DevOps Infrastructure as a code (IaC), we can automate the entire infrastructure setup. How easy does it sound? Simply put, IaC is to manage and provision the infrastructure through the code which pushes into the operational environment. The whole process flow of the development and test can deal with the complexity of the hybrid IT platform. With the NextGen Infrastructure as Code, the MSP can automate, reproduce the systems, and self-document the entire infrastructure. How easy will it be for anyone in the team, different teams, and the developers?

With the traditional IaC, flexibility, elasticity, scalability, reproducibility was a dream for the stakeholders. With the principles of NextGen IaC and AWS DevOps,  it makes it easier for collaboration and automation. It has become easier to build custom templates, configure repeatable changes, deploy as a single service or as a group. We can automate the scaling resources based on the traffic.  

The critical aspect of infrastructure is disaster recovery and backup. The traditional backup runs on fixed time intervals. Imagine, if there is a failure or network latency, it could lead to data loss that could affect the productivity and reliability of the organization. Cloud Computing has made it so simple and flexible for organizations to enhance data protection, easy deployment, and cost efficiency.

With the right back up strategy and predefined templates,  we can implement cross region backups and recovery through automation. Using reliable AWS services like S3 and Direct Connect we can sync the backup solutions at defined regular intervals. For Business Continuity Planning/Disaster Recovery (BCP/DR), we can implement CloudFormation templates for ease of use to make a highly reliable, available, and scalable or upgradeable AWS infrastructure.

25Feb 2019

Machine Learning (ML) and Artificial Intelligence(AI) are two hot catchphrases in the technology arena. ML is the subset of AI, based on the idea of providing data to machines and let them learn for themselves.

With the AWS powered ML and AI, the NextGen MSP can provide scalable infrastructure,  and deploy solutions through machine learning platforms for seamless deployment and consolidated billing.  The ML and AI positioned Enterprise Architecture for the businesses, provides faster analytics, decision making, more interaction between technology and business, reliability, and leverage for creative inexistence services.

We have frameworks for launching Infrastructure, Software, Network, and Applications. The Open Group Architecture Framework is all about the delivery part. Let’s look at the importance of Enterprise Architecture and the comparison between Traditional and NextGen Open Group Architecture.

Enterprise Architecture methodology is critical to align the concerns between IT and Business. Enterprise Architecture is the core behind any organizations productivity, agility, service, growth in revenue and cost efficiency.

The Traditional Enterprise Architecture rely upon one operating model and emphasis interdependency. For an enterprise, there will be a mix of multiple frameworks which is a long term commitment with continuous improvement.

The NextGen Enterprise Architecture methodology is a pluggable architecture comprising of dynamic compute resources, common storage platform, flexible programming, real-time support, and managing deployment. The NextGen Architecture model is a business focused model that combines both enterprise architecture and business architecture, business process management, and decision management.

The core features of NextGen Architecture is Instant customization of Network parser, application of complex rules to live network traffic, unlimited scalability and captures everything in the infrastructure, threat feeds and API

The NextGen Architecture is to communicate in real time, for that 90% of the running applications, software and servers have to be automated completely. It empowers the businesses to have a high level of flexibility, activity monitoring and actionable insights on the cost utilization. It integrates and automates solutions that enable users to plug and play experience.

The AWS powered billing and cost management ensures you pay for what you use. The AWS provides features to monitor the usage, along with the pricing calculator which could be utilized to create price estimates. The AWS has a very transparent pricing model which helps the businesses to allocate the respective budget for cloud computing.

05Sep 2018

Amazon SageMaker now supports version 1.10 in its pre-built TensorFlow containers. This makes it easier to run TensorFlow scripts, while taking advantage of the capabilities Amazon SageMaker offers, including a library of high-performance algorithms, managed and distributed training with automatic model tuning, one-click deployment, and managed hosting.

05Sep 2018

AWS CloudFormation Macros perform custom processing on CloudFormation templates from simple actions such as find-and-replace to transformation of entire templates. CloudFormation Macros use the same technology that powers AWS::Include and AWS::Serverless transforms. CloudFormation transforms help simplify template authoring by condensing the expression of AWS infrastructure as code and enabling reuse of template components.

Previously, you could use AWS::Include and AWS::Serverless transforms to process your templates that were hosted by CloudFormation. Now, you can use CloudFormation Macros to create your own custom transforms. For example, you can create common string functions for templates or define short-hand syntaxes for common CloudFormation resources. Click here to learn more about sample macros for your reference.

To learn more about CloudFormation Macros, please visit AWS CloudFormation documentation.

CloudFormation Macros are available in all AWS regions that have AWS Lambda. For a full list of AWS regions where AWS Lambda is available, please visit our Region table.

05Sep 2018

Starting today, you can enable persistent application and Windows settings for your users on AppStream 2.0. With this launch, your users’ plugins, toolbar settings, browser favorites, application connection profiles, and other settings will be saved and applied each time they start a streaming session. For example, your users can configure their plugins and toolbars for their CAD/CAM applications, and retain those settings every time they stream their application. Your users’ settings are stored in an S3 bucket you control in your AWS account.

To get started, select Stacks from the AppStream 2.0 console. Below the stacks list, choose User Settings, Application Settings Persistence, Edit. In the Application Settings Persistence dialog box, choose Enable Application Settings Persistence. To learn more about persistent application settings, see Enable Application Settings Persistence for Your AppStream 2.0 Users.

You can enable persistent application settings for your users at no additional charge in all AWS Regions where AppStream 2.0 is offered. However, you will be billed for the S3 storage used to store your user’s settings data. To use this feature, the AppStream 2.0 agent software on your image must be dated August 29, 2018 or newer. AppStream 2.0 offers pay-as-you-go pricing. Please see Amazon AppStream 2.0 Pricing for more information, and try our sample applications.

04Sep 2018

AWS Config, a service that enables you to assess, audit, and evaluate the configurations of your AWS resources, announces seven new managed rules to help you evaluate whether your AWS resource configurations comply with common best practices. This allows you to simplify compliance auditing, security analysis, change management, and operational troubleshooting.

04Sep 2018

Amazon S3 announces feature enhancements to S3 Select. S3 Select is an Amazon S3 capability designed to pull out only the data you need from an object, which can dramatically improve the performance and reduce the cost of applications that need to access data in S3.

Today, Amazon S3 Select works on objects stored in CSV and JSON format. Based on customer feedback, we’re happy to announce S3 Select support for Apache Parquet format, JSON Arrays, and BZIP2 compression for CSV and JSON objects. We are also adding support for CloudWatch Metrics for S3 Select, which lets you monitor S3 Select usage for your applications. 

04Sep 2018

Starting today, C5d instances are available in the AWS Asia Pacific (Sydney) and Asia Pacific (Tokyo) Regions. C5d instances were first introduced in May 2018 and delivers C5 instances equipped with local NVMe-based SSD block level storage physically connected to the host server. C5d instances provide high-performance block storage for applications that need access to high-speed, low latency local storage like video encoding, image manipulation and other forms of media processing. It will also benefit applications that need temporary storage of data, such as batch and log processing and applications that need caches and scratch files.

04Sep 2018

Amazon ElastiCache for Redis is now FedRAMP authorized with a Provisional Authority to Operate (P-ATO) at the High Impact Level provided by the Joint Authorization Board (JAB). United States government customers and their partners can now use the latest version of ElastiCache for Redis to process and store their FedRAMP systems, data, and mission-critical, high-impact workloads in the AWS GovCloud (US) Region, and at moderate impact level in AWS US East/West Regions.

04Sep 2018

Amazon CloudWatch Metrics for Amazon S3 now includes storage metrics for Amazon Glacier and S3 OneZone-Infrequent Access (S3 One Zone-IA) storage classes in the AWS GovCloud (US) Region. Storage that has been uploaded to S3 One Zone-IA or transitioned to S3 One Zone-IA or Amazon Glacier from S3 Standard or S3 Standard-IA storage classes using S3 Lifecycle policies will be available in CloudWatch storage metrics. These storage metrics will also include object overhead bytes applied to objects in Amazon Glacier and small objects in S3 Standard-IA and S3 One Zone-IA.

04Sep 2018

Amazon SageMaker now supports a new HTTP header for the InvokeEndpoint API action called CustomAttributes, which can be used to provide additional information about an inference request or response. Using this header, it is easy to pass custom information such as trace ID, application specific identifier or other metadata to the inference request or response. This will help customers keep track of their requests or responses for audits or tracking metrics.

03Sep 2018

AWS Glue now supports data encryption at rest for ETL jobs and development endpoints. You can configure ETL jobs and development endpoints to use AWS Key Management Service (KMS) keys to write encrypted data at rest. You can also encrypt the metadata stored in the Glue Data Catalog using keys that you manage with AWS KMS. Additionally, you can use AWS KMS keys to encrypt job bookmarks and the logs generated by crawlers and ETL jobs.

30Aug 2018

AWS WAF now supports full logging of all web requests inspected by the service. Customers can store these logs in Amazon S3 for compliance and auditing needs as well as use them for debugging and additional forensics. The logs will help customers understand why certain rules are triggered and why certain web requests are blocked. Customers can also integrate the logs with their SIEM and log analysis tools. 

30Aug 2018

The Amazon Kinesis Video Streams Producer SDK is now available for Microsoft Windows to help you stream video into AWS from sources such as webcams, USB cameras, or RTSP (network) cameras connected to your Microsoft Windows machine.

Amazon Kinesis Video Streams makes it easy to securely stream video from millions of connected devices to AWS for real-time machine learning (ML), storage, and batch-oriented processing and analytics. It also durably stores, encrypts, and indexes video data in your streams, and allows you to access your data through easy-to-use APIs.

Amazon Kinesis Video Streams provides Producer SDKs in C++ and Java that you can build, configure, and install on devices. This software makes it easier to securely and reliably stream video into AWS frame-by-frame in real-time. In addition to Mac OS, Android, Linux, and Raspbian, the C++ Producer SDK is now also available for Microsoft Windows. Developers can use the Minimal GNU for Windows (MinGW) or the Microsoft Visual Studio C++ Compiler (MSVC) to build the producer SDK from source and start streaming from cameras connected to a Microsoft Windows machine. Additionally, we have also packaged the Producer SDK GStreamer Plug-in for Windows as a Docker image so you can simply do a Docker pull and get started with streaming video in minutes. Please refer the developer documentation to learn more.

Refer to the AWS global region table for Amazon Kinesis Video Streams availability. 

Google+