csharp-nunittarafından github
NUnit best practices for standard and data-driven unit testing in .NET projects. Organize tests with [TestFixture] classes matching production code, using [Test] methods named MethodName_Scenario_ExpectedBehavior and following Arrange-Act-Assert structure Data-driven testing via [TestCase] , [TestCaseSource] , [Values] , [Range] , and [Combinatorial] attributes for inline, programmatic, and parameterized test generation Use Assert.That with constraint model ( Is.EqualTo , Contains.Item ) and...
npx skills add https://github.com/github/awesome-copilot --skill csharp-nunitNUnit Best Practices
Your goal is to help me write effective unit tests with NUnit, covering both standard and data-driven testing approaches.
Project Setup
- Use a separate test project with naming convention
[ProjectName].Tests - Reference Microsoft.NET.Test.Sdk, NUnit, and NUnit3TestAdapter packages
- Create test classes that match the classes being tested (e.g.,
CalculatorTestsforCalculator) - Use .NET SDK test commands:
dotnet testfor running tests
Test Structure
- Apply
[TestFixture]attribute to test classes - Use
[Test]attribute for test methods - Follow the Arrange-Act-Assert (AAA) pattern
- Name tests using the pattern
MethodName_Scenario_ExpectedBehavior - Use
[SetUp]and[TearDown]for per-test setup and teardown - Use
[OneTimeSetUp]and[OneTimeTearDown]for per-class setup and teardown - Use
[SetUpFixture]for assembly-level setup and teardown
Standard Tests
- Keep tests focused on a single behavior
- Avoid testing multiple behaviors in one test method
- Use clear assertions that express intent
- Include only the assertions needed to verify the test case
- Make tests independent and idempotent (can run in any order)
- Avoid test interdependencies
Data-Driven Tests
- Use
[TestCase]for inline test data - Use
[TestCaseSource]for programmatically generated test data - Use
[Values]for simple parameter combinations - Use
[ValueSource]for property or method-based data sources - Use
[Random]for random numeric test values - Use
[Range]for sequential numeric test values - Use
[Combinatorial]or[Pairwise]for combining multiple parameters
Assertions
- Use
Assert.Thatwith constraint model (preferred NUnit style) - Use constraints like
Is.EqualTo,Is.SameAs,Contains.Item - Use
Assert.AreEqualfor simple value equality (classic style) - Use
CollectionAssertfor collection comparisons - Use
StringAssertfor string-specific assertions - Use
Assert.Throws<T>orAssert.ThrowsAsync<T>to test exceptions - Use descriptive messages in assertions for clarity on failure
Mocking and Isolation
- Consider using Moq or NSubstitute alongside NUnit
- Mock dependencies to isolate units under test
- Use interfaces to facilitate mocking
- Consider using a DI container for complex test setups
Test Organization
- Group tests by feature or component
- Use categories with
[Category("CategoryName")] - Use
[Order]to control test execution order when necessary - Use
[Author("DeveloperName")]to indicate ownership - Use
[Description]to provide additional test information - Consider
[Explicit]for tests that shouldn't run automatically - Use
[Ignore("Reason")]to temporarily skip tests
github tarafından daha fazla skill
console-rendering
by github
Instructions for using the struct tag-based console rendering system in Go
acquire-codebase-knowledge
by github
Use this skill when the user explicitly asks to map, document, or onboard into an existing codebase. Trigger for prompts like "map this codebase", "document…
acreadiness-assess
by github
Run the AgentRC readiness assessment on the current repository and produce a static HTML dashboard at reports/index.html. Wraps `npx github:microsoft/agentrc…
acreadiness-generate-instructions
by github
Generate tailored AI agent instruction files via AgentRC instructions command. Produces .github/copilot-instructions.md (default, recommended for Copilot in VS…
acreadiness-policy
by github
Help the user pick, write, or apply an AgentRC policy. Policies customise readiness scoring by disabling irrelevant checks, overriding impact/level, setting…
add-educational-comments
by github
Add educational comments to code files to transform them into effective learning resources. Adapts explanation depth and tone to three configurable knowledge levels: beginner, intermediate, and advanced Automatically requests a file if none is provided, with numbered list matching for quick selection Expands files by up to 125% using educational comments only (hard limit: 400 new lines; 300 for files over 1,000 lines) Preserves file encoding, indentation style, syntax correctness, and...
adobe-illustrator-scripting
by github
Write, debug, and optimize Adobe Illustrator automation scripts using ExtendScript (JavaScript/JSX). Use when creating or modifying scripts that manipulate…
agent-governance
by github
Declarative policies, intent classification, and audit trails for controlling AI agent tool access and behavior. Composable governance policies define allowed/blocked tools, content filters, rate limits, and approval requirements — stored as configuration, not code Semantic intent classification detects dangerous prompts (data exfiltration, privilege escalation, prompt injection) before tool execution using pattern-based signals Tool-level governance decorator enforces policies at function...