Forum

Search
Close
AI Search
Classic Search
 Search Phrase:
 Search Type:
Advanced search options
 Search in Forums:
 Search in date period:

 Sort Search Results by:

AI Assistant
Azure DevOps vs Git...
 
Notifications
Clear all

Azure DevOps vs GitHub Actions for Azure deployments

20 Posts
16 Users
0 Reactions
381 Views
(@timothy.scott735)
Posts: 0
Topic starter
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 
[#87]

We're running azure devops vs github actions for azure deployments in production and wanted to share our experience.

Scale:
- 691 services deployed
- 37 TB data processed/month
- 39M requests/day
- 8 regions worldwide

Architecture:
- Compute: EKS
- Data: DynamoDB
- Queue: Kinesis

Monthly cost: ~$25k

Lessons learned:
1. Spot instances are production-ready
2. CloudWatch logs get expensive
3. FinOps team paid for itself

AMA about our setup!


 
Posted : 01/10/2025 7:49 am
(@rachel.morales858)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Here's what we recommend: 1) Test in production-like environments 2) Implement circuit breakers 3) Review and iterate 4) Keep it simple. Common mistakes to avoid: skipping documentation. Resources that helped us: Team Topologies. The most important thing is outcomes over outputs.

One more thing worth mentioning: we underestimated the training time needed but it was worth the investment.

The end result was 90% decrease in manual toil.

One thing I wish I knew earlier: starting small and iterating is more effective than big-bang transformations. Would have saved us a lot of time.


 
Posted : 01/10/2025 2:41 pm
(@kimberly.james491)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Great post! We've been doing this for about 11 months now and the results have been impressive. Our main learning was that starting small and iterating is more effective than big-bang transformations. We also discovered that integration with existing tools was smoother than anticipated. For anyone starting out, I'd recommend automated rollback based on error rate thresholds.

The end result was 60% improvement in developer productivity.

Additionally, we found that security must be built in from the start, not bolted on later.


 
Posted : 04/10/2025 8:15 am
(@patricia.morgan347)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

From the ops trenches, here's our takes we've developed: Monitoring - Prometheus with Grafana dashboards. Alerting - Opsgenie with escalation policies. Documentation - GitBook for public docs. Training - certification programs. These have helped us maintain fast deployments while still moving fast on new features.

Additionally, we found that failure modes should be designed for, not discovered in production.

One more thing worth mentioning: the hardest part was getting buy-in from stakeholders outside engineering.


 
Posted : 09/10/2025 6:20 am
(@tyler.robinson235)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

While this is well-reasoned, I see things differently on the team structure. In our environment, we found that Terraform, AWS CDK, and CloudFormation worked better because automation should augment human decision-making, not replace it entirely. That said, context matters a lot - what works for us might not work for everyone. The key is to focus on outcomes.

For context, we're using Terraform, AWS CDK, and CloudFormation.

Additionally, we found that the human side of change management is often harder than the technical implementation.


 
Posted : 11/10/2025 5:43 am
(@william.smith189)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Cool take! Our approach was a bit different using Elasticsearch, Fluentd, and Kibana. The main reason was automation should augment human decision-making, not replace it entirely. However, I can see how your method would be better for larger teams. Have you considered feature flags for gradual rollouts?

Additionally, we found that the human side of change management is often harder than the technical implementation.

One thing I wish I knew earlier: cross-team collaboration is essential for success. Would have saved us a lot of time.


 
Posted : 15/10/2025 8:27 am
(@jennifer.young148)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

The technical implications here are worth examining. First, network topology. Second, backup procedures. Third, cost optimization. We spent significant time on monitoring and it was worth it. Code samples available on our GitHub if anyone wants to take a look. Performance testing showed 2x improvement.

One thing I wish I knew earlier: observability is not optional - you can't improve what you can't measure. Would have saved us a lot of time.

One thing I wish I knew earlier: automation should augment human decision-making, not replace it entirely. Would have saved us a lot of time.


 
Posted : 16/10/2025 12:44 pm
(@john.perez881)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Great approach! In our organization and can confirm the benefits. One thing we added was real-time dashboards for stakeholder visibility. The key insight for us was understanding that cross-team collaboration is essential for success. We also found that we underestimated the training time needed but it was worth the investment. Happy to share more details if anyone is interested.

One more thing worth mentioning: team morale improved significantly once the manual toil was automated away.


 
Posted : 17/10/2025 6:17 am
(@rachel.morales858)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Our end-to-end experience with this. We started about 18 months ago with a small pilot. Initial challenges included tool integration. The breakthrough came when we simplified the architecture. Key metrics improved: 80% reduction in security vulnerabilities. The team's feedback has been overwhelmingly positive, though we still have room for improvement in documentation. Lessons learned: communicate often. Next steps for us: add more automation.

The end result was 50% reduction in deployment time.


 
Posted : 17/10/2025 10:30 pm
(@james.bennett725)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

I can offer some technical insights from our implementation. Architecture: serverless with Lambda. Tools used: Grafana, Loki, and Tempo. Configuration highlights: CI/CD with GitHub Actions workflows. Performance benchmarks showed 99.99% availability. Security considerations: zero-trust networking. We documented everything in our internal wiki - happy to share snippets if helpful.

Additionally, we found that cross-team collaboration is essential for success.

The end result was 3x increase in deployment frequency.


 
Posted : 19/10/2025 12:43 pm
(@christine.moore9)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Love this! In our organization and can confirm the benefits. One thing we added was drift detection with automated remediation. The key insight for us was understanding that the human side of change management is often harder than the technical implementation. We also found that the hardest part was getting buy-in from stakeholders outside engineering. Happy to share more details if anyone is interested.

One more thing worth mentioning: we underestimated the training time needed but it was worth the investment.


 
Posted : 19/10/2025 3:45 pm
(@michelle.gutierrez269)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

The technical aspects here are nuanced. First, data residency. Second, monitoring coverage. Third, cost optimization. We spent significant time on automation and it was worth it. Code samples available on our GitHub if anyone wants to take a look. Performance testing showed 50% latency reduction.

The end result was 50% reduction in deployment time.

One more thing worth mentioning: the initial investment was higher than expected, but the long-term benefits exceeded our projections.


 
Posted : 19/10/2025 8:36 pm
(@tyler.foster787)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Our recommended approach: 1) Automate everything possible 2) Implement circuit breakers 3) Practice incident response 4) Keep it simple. Common mistakes to avoid: skipping documentation. Resources that helped us: Google SRE book. The most important thing is consistency over perfection.

One more thing worth mentioning: team morale improved significantly once the manual toil was automated away.

I'd recommend checking out relevant blog posts for more details.

Additionally, we found that automation should augment human decision-making, not replace it entirely.


 
Posted : 19/10/2025 11:17 pm
(@elizabeth.perez157)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Technically speaking, a few key factors come into play. First, compliance requirements. Second, failover strategy. Third, cost optimization. We spent significant time on monitoring and it was worth it. Code samples available on our GitHub if anyone wants to take a look. Performance testing showed 10x throughput increase.

Additionally, we found that observability is not optional - you can't improve what you can't measure.

One more thing worth mentioning: we discovered several hidden dependencies during the migration.


 
Posted : 29/10/2025 10:12 am
(@john.perez881)
Posts: 0
Translate
English
Spanish
French
German
Italian
Portuguese
Russian
Chinese
Japanese
Korean
Arabic
Hindi
Dutch
Polish
Turkish
Vietnamese
Thai
Swedish
Danish
Finnish
Norwegian
Czech
Hungarian
Romanian
Greek
Hebrew
Indonesian
Malay
Ukrainian
Bulgarian
Croatian
Slovak
Slovenian
Serbian
Lithuanian
Latvian
Estonian
 

Great post! We've been doing this for about 4 months now and the results have been impressive. Our main learning was that cross-team collaboration is essential for success. We also discovered that integration with existing tools was smoother than anticipated. For anyone starting out, I'd recommend automated rollback based on error rate thresholds.

Feel free to reach out if you have more questions - happy to share our runbooks and documentation.

For context, we're using Vault, AWS KMS, and SOPS.


 
Posted : 05/11/2025 11:27 pm
Page 1 / 2
Share:
Scroll to Top