合肥生活安徽新闻合肥交通合肥房产生活服务合肥教育合肥招聘合肥旅游文化艺术合肥美食合肥地图合肥社保合肥医院企业服务合肥法律

XJCO3811代写、C++程序设计代做

时间:2023-12-20  来源:合肥网hfw.cc  作者:hfw.cc 我要纠错


SWJTU-Leeds Joint School: assessment brief

Module title Computer Graphics

Module code XJCO3811

Assignment title

This coursework covers the projective 3D graphics

pipeline, which is fundamental to real-time rendering

applications.

Assignment type

and description Programming assignment

Rationale

The coursework tests the student’s ability create an

interactive 3D rendering application, with custom objects

and interactions.

Page limit and

guidance

Report: 10 A4 pages with 2cm or larger margins, 10pt

font size (including figures). You are allowed to use a

double-column layout. Code: no limit. Please read

the submission instructions carefully!

Weighting 50%

Submission deadline

 2023-12-21 10:00 (Leeds Time)

2023-12-21 18:00 (Beijing Time)

Submission method Gradescope: code and report

Feedback provision Written notes

Learning outcomes

assessed

Understand, describe and utilize the projective 3D graphics pipeline and its modern implementation to create 2D

imagesfrom 3D models; Understand, describe, apply and

evaluate fundamental algorithms and methods from the

field of computer graphics, commonly used for projective

graphics applications such as 3D games.

Module lead Markus Billeter

i

1. Assignment guidance

In this second coursework, you are tasked with developing an OpenGL-based renderer

in groups of three students. You will develop an application that combines a terrain

based on real-world elevation data with pre-made models loaded from files and models

programmatically created by you.

Before starting your work, please study the coursework document in its entirety. Pay

special attention to the requirements and submission information. Plan your work. It

is likely better to focus on a subset of tasks and commit to these fully than to attempt

everything in a superficial way.

2. Assessment tasks

Please see detailed instructions in the document following the standardized assessment

brief (pages i-v). The work is split into several tasks, accounting in total for 50% of

the module grade.

3. General guidance and study support

Please refer to materials handed out during the module, including the tutorial-style

exercises for the 3D graphics pipeline.

Support is primarily provided during scheduled lab hours. Support for general issues

may be provided through the module’s “Teams” channel. Do not expect answers

outside of scheduled hours. Do not post specific issues relating to your code in the

public Teams channels. Do not crosspost across multiple channels.

4. Assessment criteria and marking process

Submissions take place through Gradescope. Valid submissions will be marked

primarily based on the report and secondarily based on the submitted code. See

following sections for details on submission requirements and on requirements on

the report. Marks and feedback will be provided through Minerva (and not through

Gradescope - Gradescope is only used for submissions!).

5. Submission requirements

Your coursework will be graded once you have

(a) Submitted required files (code and report) through Gradescope.

(b) If deemed necessary, participated in an extended interview with the instructor(s)

where you explain your submission in detail.

Your submission will consist of source code and a report (<=10 pages). The report is

the basis for assessment. The source code is supporting evidence for assertions made

in the report.

Submissions are made through Gradescope (do not send your solutions by email!).

You can use any of Gradescope’s mechanisms for uploading the complete solution

ii

and report. In particular, Gradescope accepts .zip archives (you should see the

contents of them when uploading to Gradescope). Do not use other archive formats

(Gradescope must be able to unpack them!). Gradescope will run preliminary checks

on your submission and indicate whether it is considered a valid submission.

The source code must compile and run as submitted on the standard SoC machines

found in the UG teaching lab (2.05 in Bragg). Your code must compile cleanly, i.e.,

it should not produce any warnings. If there are singular warnings that you cannot

resolve or believe are in error, you must list these in your report and provide an

explanation of what the warning means and why it is acceptable in your case. This is

not applicable for easily fixed problems and other bulk warnings (for example, type

conversions) – you are always expected to correct the underlying issues for such. Do

not change the warning level defined in the handed-out code. Disabling individual

warnings through various means will still require documenting them in the report.

In general, you must not use custom extensions or vendor-specific features in your

application (foe example, NVIDIA-specific OpenGL exensions). If you wish to explore

a specific advanced feature that requires such, please conctact the module leader

ahead of time and get their explicit permission. Contact the module leader by email

for this, explaining what feature you require, what you aim to do (and why that

specific feature is required for that).

Your submission must not include any “extra” files that are not required to build

or run your submission (aside from the report). In particular, you must not include

build artifacts (e.g. final binaries, .o files, ...), temporary files generated by your IDE

or other tools (e.g. .vs directory and contents) or files used by version control (e.g.

.git directory and related files). Note that some of these files may be hidden by

default, but they are almost always visible when inspecting the archive with various

tools. Do not submit unused code (e.g. created for testing). Submitting unnecessary

files may result in a deduction of marks.

While you are encouraged to use version control software/source code management

software (such as git or subversion), you must not make your solutions publicly

available. In particular, if you wish to use Github, you must use a private repository.

Members of your group should be the only users with access to that repository.

6. Presentation and referencing

Your report must be a single PDF file called report.pdf. In the report, you must list

all tasksthat you have attempted. Include screenshots for each task unless otherwise

noted in the task description! You may refer to your code in the descriptions, but

descriptions that just say “see source code” are not sufficient. Do not reproduce bulk

code in your report. If you wish to highlight a particularly clever method, a short

snippet of code is acceptable. Never show screenshots/images of code - if you wish to

include code, make sure it is rendered as text in the PDF using appropriate formatting

and layout. Screenshots must be of good quality (keep the resolution at 1280×720

iii

or higher, but scale them down in the PDF). Don’t compress the screenshots overly

much (e.g., visible compression artifacts).

Apply good report writing practices. Structure your report appropriately. Use whole

English sentences. Use appropriate grammar, punctuation and spelling. Provide figure

captions to figures/screenshots, explaining what the figure/screenshot is showing and

what the reader should pay attention to. Refer to figures from your main text. Cite

external references appropriately.

Furthermore, the UoL standard practices apply:

The quality of written English will be assessed in this work. As a minimum, you must

ensure:

• Paragraphs are used

• There are links between and within paragraphs although these may be ineffective

at times

• There are (at least) attempts at referencing

• Word choice and grammar do not seriously undermine the meaning and comprehensibility of the argument

• Word choice and grammar are generally appropriate to an academic text

These are pass/ fail criteria. So irrespective of marks awarded elsewhere, if you do

not meet these criteria you will fail overall.

7. Academic misconduct and plagiarism

You are encouraged to research solutions and use third-party resources. If you find

such, you must provide a reference to them in your report (include information about

the source and original author(s)). Never “copy-paste” code from elsewhere – all code

must be written yourself. If the solution is based on third party code, make sure to

indicate this in comments surrounding your implementation in your code, in addition

to including a reference in your report.

If you wish to use a custom third party library, you must clear this with the instructors

for XJCO3811 in writing well ahead of your submission. You must provide a clear

reason for using this library in your request. Third party libraries may not be used to

solve substantial portions of any of the tasks and cannot be used to replace elements

provided with the base code.

Furthermore, the UoL standard practices apply:

Academic integrity means engaging in good academic practice. This involves essential

academic skills, such as keeping track of where you find ideas and information and

referencing these accurately in your work.

iv

By submitting this assignment you are confirming that the work is a true expression

of your own work and ideas and that you have given credit to others where their work

has contributed to yours.

8. Assessment/marking criteria grid

(See separate document.)

v

Contents

1 Tasks 1

1.1 Matrix/vector functions 2

1.2 3D renderer basics ................. 2

1.3 Texturing................................ 2

1.4 Simple Instancing.................. 3

1.5 Custom model ....................... 4

1.6 Local light sources................. 4

1.7 Animation .............................. 4

1.8 Tracking cameras................... 4

1.9 Split screen ............................. 4

1.10 Particles .................................. 5

1.11 Simple 2D UI elements . 5

1.12 Measuring performance 5

2 Implementation Requirements 6

In this coursework, you will create a renderer using the (modern) OpenGL API. The coursework is to be

completed in groups. Each group can consist of one – three students. You must build on the code provided

with this coursework. In particular, the project must be buildable through the standard premake steps (see

exercises). Carefully study the submission requirements for details.

While you are encouraged to use version control software/source code management software (such as git or subversion),

you must not make your solutions publicly available. In particular, if you wish to use Github, you must use a private

repository. The members of the group should be the only users with access to that repository.

1 Tasks

The total achievable score in Coursework 2 is 50 marks. The coursework is split into the tasks listed below.

Please study all tasks and plan your work before you start coding. It is likely better to focus on a subset of

tasks and commit to these fully than to attempt everything in a superficial way.

Background: The privately held space company, AvaruusY, is preparing for lift-off. As part of their preparations, they are looking for an interactive 3D experience that lets the public explore their launch. You have

been contracted for this job! The launch is planned to take place from a sea-based launch platform in southern

Finland. AvaruusY engineers have spent the last few days furiously compiling the data that they have about

the launch site and have produced a few 3D models for your use.

Important: In the end of the report (appendix), you must include a table that lists each group member’s

individual contributions to each of the completed tasks. Document who wrote what code. Make sure that all

group members contribute substantially to the code. Prefer working together on tasks rather than distributing

them to individuals. When multiple members worked on a feature (e.g., pair programming), include rough

proportions.

XJCO3811

Coursework 2

1.1 Matrix/vector functions

4 marks

Your first task is to implement the necessary matrix and vector functions in vmlib. Minimally, this includes

the following functions in vmlib/mat44.hpp:

• Mat44f operator*( Mat44f const&, Mat44f const& )

• Vec4f operator*( Mat44f const&, Vec4f const& )

• Mat44f make_rotation_{x,y,z}( float ) (three functions)

• Mat44f make_translation( Vec3f )

• Mat44f make_perspective_projection( float, float, float, float )

These functions are quite fundamental and needed for the rest of this coursework. It is therefore important to

be able to rely on them. To help you check for the correctness, the handed-out code includes the vmlib-test

subproject. It uses the Catch2 testing library. A sample test case for the standard projection matrix is included

as an example.

You should add tests for each of the above functions. Comparing against known-good values is sufficient in

this case (but make sure your known-good values are indeed correct!). Deriving the known-good values by

hand or using a (well-known) third-party library are both acceptable. Make sure your tests are meaningful.

Note: you are required to use the vmlib types in your solution – i.e., you may not use a third party math library!

In your report: Briefly outline the tests that you have added. Explain why those tests are meaningful. (No

screenshots are required.)

1.2 3D renderer basics

8 marks

Next, set up a 3D rendering application using modern shader-based OpenGL in the main subproject. You

may start with the skeleton provided in main/main.cpp. Refer to the OpenGL exercises from the module as

necessary.

The base code includes a Wavefront OBJ file, assets/parlahti.obj and assets/parlahti.mtl. The

mesh shows a small part of Finland, near the Parlahti region with the islands Heina¨saari, Sommaro¨n and

La˚ ngo¨n (saari is island in Finnish; o¨is island in Swedish). The region is 6km by 6km and derived from a digital

elevation model with 2m resolution. The elevation is somewhat exaggerated in the model, to make the terrain

more visible. The original elevation data was obtained from Maanmittauslaitos, the national land survey

of Finland. You can download other parts of Finland from their homepage. The datasets are subject to the

Creative Commons Attribution 4.0 license.

You should set up a program that can load this Wavefront OBJ file and display it.

Use a perspective projection for rendering. You must use the make_perspective_projection function from

vmlib to create the projection matrix. Make sure the window is resizable.

Implement a camera with which you can explore the scene using the mouse and keyboard. Aim for a firstperson style 3D camera (WSAD+EQ to control position, mouse to look around, shift to speed up, and ctrl

to slow down). Camera controls should be frame-rate independent. Implement this using the callback-style

input from GLFW – do not use polling for mouse and keyboard input (so, no glfwGetKey and similar).

Use the simplified directional light model from Exercise G.5, i.e., with only a ambient and diffuse component.

For now, use the light direction (0, 1, −1) (but remember to normalize it!).

Refer to Figure 1 for some sample screenshots of what this may look like.

In your report: Report the values for GL_RENDERER, GL_VENDOR and GL_VERSION for your test computer(s).

Include several (3+) screenshots of the scene. They must be significantly different (=different views) from the

ones shown in Figure 1.

1.3 Texturing

4 marks

In addition to the elevation data, Maanmittauslaitos also provides orthophotos. These are photos taken with a

top down view (for example, from an airplane). We can use this orthophoto as a texture for the terrain mesh.

The Wavefront OBJ includes all the necessary data. Update your renderer to draw the mesh with a texture.

Combine the texture with the simple lighting from Section 1.2.

XJCO3811 - Coursework 2 3

Figure 1: Screenshots showing the Parlahti mesh, with a simple “n dot l” lighting model. The simple lighting model helps

us see the geometry clearly.

Figure 2: Screenshots showing the Parlahti model with the orthophoto-based texture. Note that it does not include any

lighting. (You should however combine the texture with lighting in your implementation!)

See Figure 2 for a textured rendering (no lighting).

In your report: Include several (3+) screenshots with texturing. They must be significantly different (=different

views) from the ones shown in Figure 2.

1.4 Simple Instancing

4 marks

“Why build one when you can have two at twice the price?”, Contact (1997)

The exercise includes a second Wavefront OBJ file, assets/launchpad.obj and assets/launchpad.mtl

(Figure 3). This models the launchpad that AvaruusY is planning to use for its launches.

Load this file (once). Find two spots for the launch pad. They must be in the sea and in contact with the water

(but not fully submerged). They cannot be at the origin. They should not be immediately next to each other.

Render the object twice (without making a copy of the data), once at each location. The model does not have

a texture, but uses per-material colors. You will need to handle this somehow, for example with two different

shader programs.

In your report: Document the coordinates at which you placed the launch pads. Include screenshots of the

placement.

Figure 3: Launchpad model (launchpad.obj).

1.5 Custom model

4 marks

Create a 3D model of a space vehicle programmatically. Combine basic shapes (boxes, cylinders, ...) into a

complex object. You must use at least seven (7) basic shapes. These must be connected to each other (e.g.,

they can not float freely) and all of the parts must be visible. You must use at least three (3) different basic

shapes. Simple transformations of a shape, such as a non-uniform scaling, do not constitute a different shape

(for example, transforming a cube into a box still only counts as one shape). You should however use all

transformations (translation, rotation, scaling) when placing objects. At least one object must be placed relative

to another (e.g., you place one thing on top of another).

Make sure you generate the appropriate normals for the object, such that it is lit correctly. You do not have to

texture the object. You are free to use different colors for different parts if you wish.

The space vehicle must be placed on one of the launch pads (Section 1.4).

Note: You have some freedom in designing your “space vehicle”. It should be recognizable as a vehicle, but it

does not have to be a rocket. An alien UFO would work as well. Or a submarine ... after all, that’s a bit like a

spaceship, except that it goes the other way1

.

In your report: Document which of the launch pads you placed your space vehicle on. Include screenshots

(2+) of your vehicle.

1.6 Local light sources

3 marks

Implement the full Blinn-Phong shading model for point lights, including the standard 1/r

2 distance attenuation. Add three point lights, each with a different color, to your space vehicle (Section 1.5). You may place

them slightly away from the geometry to avoid problems with the division.

In your report: Include a screenshot of lit space vehicle and launch pad.

1.7 Animation

3 marks

Implement an animation where your space vehicle flies away. The animation should start when the user

presses the F key. Pressing R should reset the space vehicle to its original position and disable the animation.

For full marks, you should consider a curved path for the vehicle, where it slowly accelerates from a standstill.

The vehicle should rotate such that it always faces into the direction of movement. The lights from Section 1.6

should follow the space vehicle.

In your report: Include a screenshot of the vehicle mid-flight.

1.8 Tracking cameras

3 marks

Implement a feature where the user can switch to cameras that track the space vehicle in its flight (Section 1.7).

Implement two cameras modes:

• A camera that looks at the space vehicle from a fixed distance and follows it in its flight.

• A camera that is fixed on the ground and always looks towards the space vehicle, even as it flies away.

The user should be able to cycle through the different camera modes with the C key. So, pressing the key

should first switch from the freely controllable default camera to the fixed-distance camera, next to the ground

camera and then back.

In your report: Include a screenshot of each camera mode, with the vehicle in mid-flight.

1.9 Split screen

4 marks

Implement a split screen rendering mode that shows two different views at once. Let the user toggle split

screen mode by pressing the V key. Pressing C should select the camera mode for one of the views (Section 1.8).

The camera mode for the other one should cycle by pressing Shift-C. You can pick how you split the screen,

but a simple horizontal 50%-50% split is fine.

1The AvaruusY engineers might object to a car, though, since a different company already sent one of those to space.

XJCO3811 - Coursework 2 5

There are multiple possible strategies for implementing a split screen mode. You are free to pick one yourself.

However, it should not be overly wasteful (so don’t render parts that are never shown to the user). The split

screen mode should adapt to the window when resized.

In your report: Describe how you have implemented the split screen feature. Include screenshots that show

the split screen mode in action.

1.10 Particles

4 marks

Implement a particle system that simulates exhaust from your space vehicle as it flies away (Section 1.7).

For full marks, consider the following features:

• Render particles as textured point sprites or as textured camera-facing quads.

• Make sure the particles render correctly from all views. Designing the system around additive blending

will make this easier.

• Particles should be emitted in whatever small area you consider the vehicle’s engine, and should appear

to move away from the vehicle.

• Particles should have fixed lifetime after which they disappear.

Take particular care with the implementation. For example, avoid dynamic allocations at runtime during the

simulation. A CPU-based implementation is perfectly adequate.

In your report: Describe how you have implemented the particle system. Document any assumptions that you

have made and mention limitations. Discuss the efficiency/performance of the system. Include a screenshot

of the effect.

1.11 Simple 2D UI elements

4 marks

Implement the following UI elements:

• In the top left corner, display the current altitude of the space vehicle in text.

• In the bottom center, implement two buttons. One that launches the space vehicle when clicked and one

that resets the vehicle to its initial position (see Section 1.7).

Make sure the buttons stay in the right place when the window is resized.

Draw the buttons such that they their label is centered in a rectangle with a solid outline and a semi-transparent

fill. The buttons should have three distinct visual states: normal, mouse-over and pressed (mouse button held

down). The top-left status text should remain readable at all times.

You should implement this yourself, e.g., without a third party UI library. You may however use Fontstash

for rendering text (use the included assets/DroidSansMonoDotted.ttf font). You can also opt for rendering text using a font atlas texture if you prefer.

In your report: Describe your implementation. What are the steps to adding another UI element? Include

screenshots of the UI, including the different button states.

1.12 Measuring performance

5 marks

Measure the performance of your renderer. This task primarily focuses on the GPU rendering performance

(recall that the GPU will run asynchronously from the CPU). You must therefore use GL_TIMESTAMP queries

with the glQueryCounter() function. Study the documentation for the above function. Refer to OpenGL

wiki on query objects for additional information.

Use the queries to measure rendering performance. Each frame, you should measure:

• the fullrendering time (but exclude e.g., glfwSwapBuffers() and other non-rendering code).

• the time needed to render the parts from Sections 1.2, 1.4 and 1.5 individually, so as three different times.

If you implemented Section 1.9, you can measure this only for one of the views or for both individually. For

full marks, you should implement a strategy that avoids stalling/waiting when retrieving results from the

asynchronous timestamp queries. No marks will be awarded if you use third party applications to measure

the performance.

Additionally measure frame-to-frame times and the time it takes to submit the rendering commands with

OpenGL using a CPU timer such as std::chrono::high_resolution_clock.

Perform the measurements over multiple frames. Make sure you run in release mode and that you have

disabled any OpenGL debugging and removed any potentially blocking calls (e.g., glGetError()). Ideally,

try to run this on one of the newer machines in 2.05, e.g., with the NVIDIA RTX GPUs.

In your report: Describe your implementation. Document your results (table or plot) and discuss these. Compare the different timings. Are they what you would expect? Are the timings reproducible in different runs?

Do they vary as you move around? Explain your observations. (No screenshots required.)

2 Implementation Requirements

For full marks, you are expected to following the requirements/specification detailed in this document. Further, you are expected apply to good (graphics) engineering practices, including:

• Manage resources! In debug mode, you should clean up all resources and allocations. In release mode,

you may choose to “leak” some resources on exit as an optimization (which the OS will free up once the

process terminates), but in order to aid debugging, you must not do so in debug builds. Furthermore,

you should never leak run-time allocations that cause resource usage to continually increase.

• Do not unnecessarily duplicate data. Typically, any large resource should be allocated at most once and

never be duplicated. If you duplicate resources, you should make this an informed choice and give a

reason for doing so (e.g., double-buffering due to concurrent accesses, or in order to enable batching for

more efficient draw calls).

• Do not do unnecessary work per frame. Avoid allocating and deallocating data each frame; allocate upfront and re-use resources instead. Don’t do file IO per frame. Don’t repeatedly query data that doesn’t

change (for example, don’t glGetUniformLocation each frame!).

• Do not do unnecessary work in the pipeline: Don’t do computations in the fragment shader if they could

be done in the vertex shader. Don’t do computations in the vertex shader if they could be done once on

the CPU. Don’t repeat computations each frame if they can (reasonably) be done once on startup. Etc.

(This is sometimes a trade-off between memory and compute. If in doubt, motivate your choice.)

• Use appropriate data structures. For example, by default you should prefer an std::vector over a

linked list. Don’t use a shared_ptr when an unique_ptr would do. ...

• Don’t reinvent the wheel unnecessarily. For example, do use the C++ standard library’s containers,

rather than implementing your own. (The exercise requires you to implement your own vector/matrix

functions. This can be considered a special exception to this guideline.)

Wrapping up

Please double-check the requirements in and ensure that your submission conforms to these. In particular,

pay attention to file types (archive format and report format) and ensure that you have not included any

unnecessary files in the submission.

Make sure you have included the table with individual contributions in your report (see Section 1).

请加QQ:99515681 或邮箱:99515681@qq.com   WX:codehelp

 

扫一扫在手机打开当前页
  • 上一篇:代写DS 2043、SQL语言编程代做
  • 下一篇:代写CS345、代做Java,Python程序设计
  • 无相关信息
    合肥生活资讯

    合肥图文信息
    新能源捕鱼一体电鱼竿好用吗
    新能源捕鱼一体电鱼竿好用吗
    海信罗马假日洗衣机亮相AWE  复古美学与现代科技完美结合
    海信罗马假日洗衣机亮相AWE 复古美学与现代
    合肥机场巴士4号线
    合肥机场巴士4号线
    合肥机场巴士3号线
    合肥机场巴士3号线
    合肥机场巴士2号线
    合肥机场巴士2号线
    合肥机场巴士1号线
    合肥机场巴士1号线
    合肥轨道交通线路图
    合肥轨道交通线路图
    合肥地铁5号线 运营时刻表
    合肥地铁5号线 运营时刻表
  • 币安app官网下载

    关于我们 | 打赏支持 | 广告服务 | 联系我们 | 网站地图 | 免责声明 | 帮助中心 | 友情链接 |

    Copyright © 2024 hfw.cc Inc. All Rights Reserved. 合肥网 版权所有
    ICP备06013414号-3 公安备 42010502001045