- Agisoft photoscan user manual professional edition version 1.3 free

- Agisoft photoscan user manual professional edition version 1.3 free

Looking for:

Agisoft Metashape: User Manuals.Agisoft PhotoScan User Manual. Standard Edition, Version PDF Free Download 













































   

 

- Agisoft photoscan user manual professional edition version 1.3 free



 

Texture mapping modes The texture mapping mode determines how the object texture will be packed in the texture atlas. Proper texture mapping mode selection helps to obtain optimal texture packing and, consequently, better visual quality of the final model.

Generic The default mode is the Generic mapping mode; it allows to parametrize texture atlas for arbitrary geometry. No assumptions regarding the type of the scene to be processed are made; program tries to create as uniform texture as possible.

Adaptive orthophoto In the Adaptive orthophoto mapping mode the object surface is split into the flat part and vertical regions. The flat part of the surface is textured using the orthographic projection, while vertical regions are textured separately to maintain accurate texture representation in such regions.

Orthophoto In the Orthophoto mapping mode the whole object surface is textured in the orthographic projection. The Orthophoto mapping mode produces even more compact texture representation than the Adaptive orthophoto mode at the expense of texture quality in vertical regions. Single photo The Single photo mapping mode allows to generate texture from a single photo. The photo to be used for texturing can be selected from 'Texture from' list.

Keep uv The Keep uv mapping mode generates texture atlas using current texture parametrization. It can be used to rebuild texture atlas using different resolution or to generate the atlas for the model parametrized in the external software. Texture generation parameters The following parameters control various aspects of texture atlas generation: Texture from Single photo mapping mode only Specifies the photo to be used for texturing.

Available only in the Single photo mapping mode. Blending mode not used in Single photo mode Selects the way how pixel values from different photos will be combined in the final texture. Mosaic - implies two-step approach: it does blending of low frequency component for overlapping images to avoid seamline problem weighted average, weight being dependent on a number of parameters including proximity of the pixel in question to the center of the image , while high frequency component, that is in charge of picture details, is taken from a single image - the one that presents good resolution for the area of interest while the camera view is almost along the normal to the reconstructed surface in that point.

Average - uses the weighted average value of all pixels from individual photos, the weight being dependent on the same parameters that are considered for high frequency component in mosaic mode. Max Intensity - the photo which has maximum intensity of the corresponding pixel is selected. Min Intensity - the photo which has minimum intensity of the corresponding pixel is selected. Disabled - the photo to take the color value for the pixel from is chosen like the one for the high frequency component in mosaic mode.

Exporting texture to several files allows to archive greater resolution of the final model texture, while export of high resolution texture to a single file can fail due to RAM limitations. Multi-page texture atlas generation is supported for Generic mapping mode only and Keep UV option, if the imported model contains proper texture layout. However, please note that color correction process takes up quite a long time, so it is recommended to enable the setting only for the data sets that proved to present results of poor quality.

Enable hole filling This option is enabled on default since it helps to avoid salt-and-pepper effect in case of complicated surface with numerous tiny parts shading other parts of the model. Only in case of very specific tasks might it be recommended to switch the function off. Improving texture quality To improve resulting texture quality it may be reasonable to exclude poorly focused images from processing at this step.

PhotoScan suggests automatic image quality estimation feature. PhotoScan estimates image quality as a relative sharpness of the photo with respect to other images in the data set. Saving intermediate results Certain stages of 3D model reconstruction can take a long time. The full chain of operations could eventually last for hours when building a model from hundreds of photos. It is not always possible to complete all the operations in one run. PhotoScan allows to save intermediate results in a project file.

This includes mesh and texture if it was built. Note that since PhotoScan tends to generate extra dense point clouds and highly detailed polygonal models, project saving procedure can take up quite a long time. You can decrease compression level to speed up the saving process. However, please note that it will result in a larger project file.

Compression level setting can be found on the Advanced tab of the Preferences dialog available from Tools menu. You can save the project at the end of any processing stage and return to it later. To restart work simply load the corresponding file into PhotoScan.

Project files can also serve as backup files or be used to save different versions of the same model. Project files use relative paths to reference original photos.

Thus, when moving or copying the project file to another location do not forget to move or copy photographs with all the folder structure involved as well. Otherwise, PhotoScan will fail to run any operation requiring source images, although the project file including the reconstructed model will be loaded up correctly.

Alternatively, you can enable Store absolute image paths option on the Advanced tab of the Preferences dialog available from Tools menu. Exporting results PhotoScan supports export of processing results in various representations: sparse and dense point clouds, camera calibration and camera orientation data, mesh, etc. Point cloud and camera calibration data can be exported right after photo alignment is completed.

All other export options are available after the corresponding processing step. To align the model orientation with the default coordinate system use Rotate object button from the Toolbar. In some cases editing model geometry in the external software may be required. PhotoScan supports model export for editing in external software and then allows to import it back as it is described in the Editing model geometry section of the manual.

Main export commands are available from the File menu and the rest from the Export submenu of the Tools menu. Point cloud export To export sparse or dense point cloud 1. Select Export Points Browse the destination folder, choose the file type, and print in the file name. Click Save button. Indicate export parameters applicable to the selected file type. Click OK button to start export. In some cases it may be reasonable to edit point cloud before exporting it.

To read about point cloud editing refer to the Editing point cloud section of the manual. Tie points data export To export matching points 1.

Select Export Matches In the Export Matches dialog box set export parameters. Precision value sets the limit to the number of decimal digits in the tie points coordinates to be saved.

Later on, estimated camera data can be imported back to PhotoScan using Import Cameras command from the Tools menu to proceed with 3D model reconstruction procedure. Camera calibration and orientation data export To export camera calibration and camera orientation data select Export Cameras To indicate for the software that loaded images have been taken from one camera station, one should move those photos to a camera group and assign Camera Station type to it.

For information on camera groups refer to Loading photos section. Select Export - Export Panorama Select camera group which panorama should be previewed for. Choose panorama orientation in the file with the help of navigation buttons to the right of the preview window in the Export Panorama dialog. Set exporting parameters: select camera groups which panorama should be exported for and indicate export file name mask. Click OK button 6. Browse the destination folder and click Save button.

Additionally, you can set boundaries for the region of panorama to be exported using Setup boundaries section of the Export Panorama dialog. Text boxes in the first line X allow to indicate the angle in the horizontal plane and the second line Y serves for angle in the vertical plane limits. Image size option enables to control the size of the exporting file. Select Export Model In the Export Model dialog indicate export parameters applicable to the selected file type.

The texture file should be kept in the same directory as the main file describing the geometry. If the texture atlas was not built only the model geometry is exported. PhotoScan supports direct uploading of the models to Sketchfab resource. To publish your model online use Upload Model Capture View command available from the context menu shown on right button click in the Model or Ortho view.

PhotoScan supports direct uploading of the 3D models to Sketchfab resource. Improving camera alignment results Camera calibration Calibration groups While carrying out photo alignment PhotoScan estimates both internal and external camera orientation parameters, including nonlinear radial distortions. For the estimation to be successful it is crucial to apply the estimation procedure separately to photos taken with different cameras.

All the actions described below could and should be applied or not applied to each calibration group individually.

Calibration groups can be rearranged manually. To create a new calibration group 1. Select Camera Calibration In the Camera Calibration dialog box, select photos to be arranged in a new group. In the right-click context menu choose Create Group command. A new group will be created and depicted on the left-hand part of the Camera Calibration dialog box. To move photos from one group to another 1. In the Camera Calibration dialog box choose the source group on the left-hand part of the dialog.

Select photos to be moved and drag them to the target group on the left-hand part of the Camera Calibration dialog box. To place each photo into a separate group you can use Split Groups command available at the right button click on a calibration group name in the left-hand part of the Camera Calibration dialog Camera types PhotoScan supports two major types of camera: frame camera and fisheye camera.

Camera type can be set in Camera Calibration dialog box available from Tools menu. Frame camera. If the source data within a calibration group was shot with a frame camera, for successful estimation of camera orientation parameters the information on approximate focal length pix is required.

Obviously, to calculate focal length value in pixel it is enough to know focal length in mm along with the sensor pixel size in mm. Normally this data is extracted automatically from the EXIF metadata.

Frame camera with Fisheye lens. If extra wide lenses were used to get the source data, standard PhotoScan camera model will not allow to estimate camera parameters successfully. Fisheye camera type setting will initialize implementation of a different camera model to fit ultra-wide lens distortions.

However, if the initial guess values differ significantly from the actual focal length, it is likely to lead to failure of the alignment process. So, if photos do not contain EXIF metadata, it is preferable to specify focal length mm and sensor pixel size mm manually.

It can be done in Camera Calibration dialog box available from Tools menu. Generally, this data is indicated in camera specification or can be received from some online source. To indicate to the program that camera orientation parameters should be estimated based on the focal length and pixel size information, it is necessary to set the Type parameter on the Initial tab to Auto value. Camera calibration parameters Once you have tried to run the estimation procedure and got poor results, you can improve them thanks to the additional data on calibration parameters.

To specify camera calibration parameters 1. Select calibration group, which needs reestimation of camera orientation parameters on the left side of the Camera Calibration dialog box. In the Camera Calibration dialog box, select Initial tab. Modify the calibration parameters displayed in the corresponding edit boxes.

Set the Type to the Precalibrated value. Repeat to every calibration group where applicable. Click OK button to set the calibration. Initial calibration data will be adjusted during the Align Photos processing step. Once Align Photos processing step is finished adjusted calibration data will be displayed on the Adjusted tab of the Camera Calibration dialog box.

If very precise calibration data is available, to protect it from recalculation one should check Fix calibration box. In this case initial calibration data will not be changed during Align Photos process. Adjusted camera calibration data can be saved to file using Save button on the Adjusted tab of the Camera Calibration dialog box. Estimated camera distortions can be seen on the distortion plot available from context menu of a camera group in the Camera Calibration dialog.

In addition, residuals graph the second tab of the same Distortion Plot dialog allows to evaluate how adequately the camera is described with the applied mathematical model.

Note that residuals are averaged per cell of an image and then across all the images in a camera group. Calibration parameters list f Focal length measured in pixels. Optimization Optimization of camera alignment During photo alignment step PhotoScan automatically finds tie points and estimates intrinsic and extrinsic camera parameters. However, the accuracy of the estimates depends on many factors, like overlap between the neighbouring photos, as well as on the shape of the object surface.

Thus, it is recommended to inspect alignment results in order to delete tie points with too large reprojection error if any. Please refer to Editing point cloud section for information on point cloud editing.

Once the set of tie points has been edited, it is necessary to run optimization procedure to reestimate intrinsic and extrinsic camera parameters. Optimization procedure calculates intrinsic and extrinsic camera parameters based on the tie points left after editing procedure. Providing that outliers have been removed, the estimates will be more accurate.

In addition, this step involves estimation of a number of intrinsic camera parameters which are fixed at the alignment step: aspect, skew; and distortion parameters p3, p4, k4. To optimize camera alignment 1. Choose Optimize Cameras In Optimize Camera Alignment dialog box check camera parameters to be optimized. Click OK button to start optimization. After optimization is complete, estimated intrinsic camera parameters can be inspected on the Adjusted tab of the Camera Calibration dialog available from the Tools menu.

You will have to rebuild the model geometry after optimization. Editing Using masks Overview Masks are used in PhotoScan to specify the areas on the photos which can otherwise be confusing to the program or lead to incorrect reconstruction results.

Thus, the objects on the masked parts of the photos are not taken into account while estimating camera positions. This is important in the setups, where the object of interest is not static with respect to the scene, like when using a turn table to capture the photos.

Masking may be also useful when the object of interest occupies only a small part of the photo. In this case a small number of useful matches can be filtered out mistakenly as a noise among a much greater number of matches between background objects. Building dense point cloud While building dense point cloud, masked areas are not used in the depth maps computation process.

Masking can be used to reduce the resulting dense cloud complexity, by eliminating the areas on the photos that are not of interest. Masked areas are always excluded from processing during dense point cloud and texture generation stages, including Tiled Model generation process.

Let's take for instance a set of photos of some object. Along with an object itself on each photo some background areas are present. These areas may be useful for more precise camera positioning, so it is better to use them while aligning the photos. However, impact of these areas at the building dense point cloud is exactly opposite: the resulting model will contain object of interest and its background.

Background geometry will "consume" some part of mesh polygons that could be otherwise used for modeling the main object. Setting the masks for such background areas allows to avoid this problem and increases the precision and quality of geometry reconstruction. Masking areas on the photos that are occluded by outliers or obstacles helps to prevent the "ghosting" effect on the resulting texture atlas.

Loading masks Masks can be loaded from external sources, as well as generated automatically from background images if such data is available.

To import masks 1. Select Import Masks In the Import Mask dialog select suitable parameters. When generating masks from separate or background images, the folder selection dialog will appear. Browse to the folder containing corresponding images and select it. The following parameters can be specified during mask import: Method Specifies the source of the mask data.

From Alpha - load masks from alpha channel of the source photos. From File - load masks from separate images. From Background - generate masks from background photos. From Model - generate masks based on reconstructed model. Operation Specifies the action to be done in case a second mask is imported for the photo.

Replacement - new mask will be loaded and stored instead of the original one. Union - two masks will be united and stored. Intersection - the intersection of the two masks will be stored as a new mask for the photo. Difference - only the difference between two masks will be stored as a new mask for the photo. Filename template not used in From alpha mode Specifies the file name template used to generate mask file names. This template can contain special tokens, that will be substituted by corresponding data for each photo being processed.

Tolerance From Background method only Specifies the tolerance threshold used for background differencing. Tolerance value should be set according to the color separation between foreground and background pixels. For larger separation higher tolerance values can be used.

Apply to Specifies whether masks should be imported for the currently opened photo, active chunk or entire Workspace. All cameras - load masks for active chunk. Entire workspace - load masks for all chunks in the project. Selected cameras - load mask for the currently checked cameras if any. Current photo - load mask for the currently opened photo if any. Editing masks Modification of the current mask is performed by adding or subtracting selections.

A selection is created with one of the supported selection tools and is not incorporated in the current mask until it is merged with a mask using Add Selection or Subtract Selection operations.

To edit the mask 1. The photo will be opened in the main window. The existing mask will be displayed as a shaded region on the photo. Select the desired selection tool and generate a selection. Click on Add Selection toolbar button to add current selection to the mask, or Subtract Selection to subtract the selection from the mask. Invert Selection button allows to invert current selection prior to adding or subtracting it from the mask. The following tools can be used for creating selections: Rectangle selection tool Rectangle selection tool is used to select large areas or to clean up the mask after other selection tools were applied.

Intelligent scissors tool Intelligent scissors is used to generate a selection by specifying its boundary. The boundary is formed by selecting a sequence of vertices with a mouse, which are automatically connected with segments. To enable snapping, hold Ctrl key while selecting the next vertex. To complete the selection, the boundary should be closed by clicking on the first boundary vertex. Intelligent paint tool Intelligent paint tool is used to "paint" a selection by the mouse, continuously adding small image regions, bounded by object boundaries.

Magic wand tool Magic Wand tool is used to select uniform areas of the image. To make a selection with a Magic Wand tool, click inside the region to be selected. The range of pixel colors selected by Magic Wand is controlled by the tolerance value. At lower tolerance values the tool selects fewer colors similar to the pixel you click with the Magic Wand tool.

Higher value broadens the range of colors selected. A mask can be inverted using Invert Mask command from the Photo menu. The command is active in Photo View only. Alternatively, you can invert masks either for selected cameras or for all cameras in a chunk using Invert Masks The masks are generated individually for each image. If some object should be masked out, it should be masked out on all photos, where that object appears. Saving masks Created masks can be also saved for external editing or storage.

To export masks 1. Select Export Masks In the Export Mask dialog select suitable parameters. Browse to the folder where the masks should be saved and select it. The following parameters can be specified during mask export: Export masks for Specifies whether masks should be exported for the currently opened photo, active chunk or entire Workspace.

Current photo - save mask for the currently opened photo if any. Active chunk - save masks for active chunk. Entire workspace - save masks for all chunks in the project. Single channel mask image - generates single channel black and white mask images. Image with alpha channel - generates color images from source photos combined with mask data in alpha channel. Mask file names Specifies the file name template used to generate mask file names.

Mask file names parameter will not be used in this case. Filtering points based on specified criterion In some cases it may be useful to find out where the points with high reprojection error are located within the sparse cloud, or remove points representing high amount of noise. Point cloud filtering helps to select such points, which usually are supposed to be removed. It is also typical for false matches.

Removing such points can improve accuracy of the subsequent optimization step. Reconstruction uncertainty High reconstruction uncertainty is typical for points, reconstructed from nearby photos with small baseline. Such points can noticeably deviate from the object surface, introducing noise in the point cloud. While removal of such points should not affect the accuracy of optimization, it may be useful to remove them before building geometry in Point Cloud mode or for better visual appearance of the point cloud.

Image count PhotoScan reconstruct all the points that are visible at least on two photos. However, points that are visible only on two photos are likely to be located with poor accuracy. Image count filtering enables to remove such unreliable points from the cloud.

Projection Accuracy This criterion allows to filter out points which projections were relatively poorer localised due to their bigger size. To remove points based on specified criterion 1. Switch to Point Cloud view mode using Point Cloud toolbar button. Select Gradual Selection In the Gradual Selection dialog box specify the criterion to be used for filtering.

Adjust the threshold level using the slider. You can observe how the selection changes while dragging the slider. Click OK button to finalize the selection. To remove selected points use Delete Selection command from the Edit menu or click Delete Selection toolbar button or simply press Del button on the keyboard.

Filtering points based on applied masks To remove points based on applied masks 1. Switch to Dense Cloud view mode using Dense Cloud toolbar button. Choose Select Masked Points In the Select Masked Points dialog box indicate the photos whose masks to be taken into account. Adjust the edge softness level using the slider. Click OK button to run the selection procedure. Filtering points based on points colors To remove points based on points colors 1.

Choose Select Points by Color In the Select Points by Color dialog box the color to be used as the criterion. Adjust the tolerance level using the slider. Tie point per photo limit Tie point limit parameter could be adjusted before Align photos procedure. The number indicates the upper limit for matching points for every image. Using zero value doesn't apply any tie-point filtering. The number of tie points can also be reduced after the alignment process with Tie Points - Thin Point Cloud command available from Tools menu.

Manual points removal Incorrect points can be also removed manually. To remove points from a point cloud manually 1. To add new points to the current selection hold the Ctrl key during selection of additional points. To remove some points from the current selection hold the Shift key during selection of points to be removed. To delete selected points click the Delete Selection toolbar button or select Delete Selection command from the Edit menu.

To crop selection to the selected points click the Crop Selection toolbar button or select Crop Selection command from the Edit menu. PhotoScan allows to export mesh and then import it back for this purpose. Decimation tool Decimation is a tool used to decrease the geometric resolution of the model by replacing high resolution mesh with a lower resolution one, which is still capable of representing the object geometry with high accuracy.

PhotoScan tends to produce 3D models with excessive geometry resolution, so mesh decimation is usually a desirable step after geometry computation. Highly detailed models may contain hundreds of thousands polygons. While it is acceptable to work with such a complex models in 3D editor tools, in most conventional tools like Adobe Reader or Google Earth high complexity of 3D models may noticeably decrease application performance.

High complexity also results in longer time required to build texture and to export model in pdf file format. In some cases it is desirable to keep as much geometry details as possible like it is needed for scientific and archive purposes. However, if there are no special requirements it is recommended to decimate the model down to - polygons for exporting in PDF, and to or even less for displaying in Google Earth and alike tools.

To decimate 3D model 1. Select Decimate Mesh In the Decimate Mesh dialog box specify the target number of polygons, which should remain in the final model. Click on the OK button to start decimation. To cancel processing click on the Cancel button. You will have to rebuild texture atlas after decimation is complete. Close Holes tool Close Holes tool provides possibility to repair your model if the reconstruction procedure resulted in a mesh with several holes, due to insufficient image overlap for example.

Close holes tool enables to close void areas on the model substituting photogrammetric reconstruction with extrapolation data. It is possible to control an acceptable level of accuracy indicating the maximum size of a hole to be covered with extrapolated data. To close holes in a 3D model 1. Select Close Holes In the Close Holes dialog box indicate the maximum size of a hole to be covered with the slider.

Click on the OK button to start the procedure. Polygon filtering on specified criterion In some cases reconstructed geometry may contain the cloud of small isolated mesh fragments surrounding the "main" model or big unwanted polygons. Mesh filtering based on different criteria helps to select polygons, which usually are supposed to be removed.

PhotoScan supports the following criteria for face filtering: Connected component size This filtering criteria allows to select isolated fragments with a certain number of polygons. The number of polygons in all isolated components to be selected is set with a slider and is indicated in relation to the number of polygons in the whole model.

The components are ranged in size, so that the selection proceeds from the smallest component to the largest one. Polygon size This filtering criteria allows to select polygons up to a certain size.

The size of the polygons to be selected is set with a slider and is indicated in relation to the size of the whole model. This function can be useful, for example, in case the geometry was reconstructed in Smooth type and there is a need to remove extra polygons automatically added by PhotoScan to fill the gaps; these polygons are often of a larger size that the rest.

To remove small isolated mesh fragments 1. In the Gradual Selection dialog box select Connected component size criterion. Select the size of isolated components to be removed using the slider. To remove the selected components use Delete Selection command from the Edit menu or click Delete Selection toolbar button or simply press Del button on the keyboard.

To remove large polygons 1. In the Gradual Selection dialog box select Polygon size criterion. Select the size of polygons to be removed using the slider. Note that PhotoScan always selects the fragments starting from the smallest ones.

If the model contains only one component the selection will be empty. Manual face removal Unnecessary and excessive sections of model geometry can be also removed manually. To remove part of the mesh polygons manually 1. Select rectangle, circle or free-form selection tool using Rectangle Selection, Circle Selection or Free-Form Selection toolbar buttons.

Make the selection using the mouse. To add new polygons to the current selection hold the Ctrl key during selection of additional polygons. To remove some polygons from the current selection hold the Shift key during selection of polygons to be excluded. To delete selected polygons click the Delete Selection toolbar button or use Delete Selection command from the Edit menu.

To crop selection to the selected polygons click the Crop Selection toolbar button or use Crop Selection command from the Edit menu. To grow or shrink current selection 1. To grow current selection press PageUp key in the selection mode. To grow selection by even a larger amount, press PageUp while holding Shift key pressed. To shrink current selection press PageDown key in the selection mode.

To shrink selection by even a larger amount, press PageDown while holding Shift key pressed. Fixing mesh topology PhotoScan is capable of basic mesh topology fixing. To fix mesh topology 1. Select View Mesh Statistics In the Mesh Statistics dialog box you can inspect mesh parameters. If there are any topological problems, Fix Topology button will be active and can be clicked to solve the problems.

Editing mesh in the external program To export mesh for editing in the external program 1. In the Save As dialog box, specify the desired mesh format in the Save as type combo box. Select the file name to be used for the model and click Save button. In the opened dialog box specify additional parameters specific to the selected file format. To import edited mesh 1. Select Import Mesh In the Open dialog box, browse to the file with the edited model and click Open.

Please make sure to select one of these file formats when exporting model from the external 3D editor. Automation Using chunks When working with typical data sets, automation of general processing workflow allows to perform routine operations efficiently.

PhotoScan allows to assign several processing steps to be run one by one without user intervention thanks to Batch Processing feature. Manual user intervention can be minimized even further due to 'multiple chunk project' concept, each chunk to include one typical data set. For a project with several chunks of the same nature, common operations available in Batch Processing dialog are applied to each selected chunk individually, thus allowing to set several data sets for automatic processing following predefined workflow pattern.

In addition, multiple chunk project could be useful when it turns out to be hard or even impossible to generate a 3D model of the whole scene in one go. This could happen, for instance, if the total amount of photographs is too large to be processed at a time. To overcome this difficulty PhotoScan offers a possibility to split the set of photos into several separate chunks within the same project.

Alignment of photos, building dense point cloud, building mesh, and forming texture atlas operations can be performed for each chunk separately and then resulting 3D models can be combined together. Working with chunks is not more difficult than using PhotoScan following the general workflow. In fact, in PhotoScan always exists at least one active chunk and all the 3D model processing workflow operations are applied to this chunk.

To work with several chunks you need to know how to create chunks and how to combine resulting 3D models from separate chunks into one model. Creating a chunk To create new chunk click on the Add Chunk toolbar button on the Workspace pane or select Add Chunk command from the Workspace context menu available by right-clicking on the root element on the Workspace pane.

After the chunk is created you may load photos in it, align them, generate dense point cloud, generate mesh surface model, build texture atlas, export the models at any stage and so on. The models in the chunks are not linked with each other. The list of all the chunks created in the current project is displayed in the Workspace pane along with flags reflecting their status.

The following flags can appear next to the chunk name: R Referenced Will appear when two or more chunks are aligned with each other.

To move photos from one chunk to another simply select them in the list of photos on the Workspace pane, and then drag and drop to the target chunk. Working with chunks All operations within the chunk are carried out following the common workflow: loading photographs, aligning them, generating dense point cloud, building mesh, building texture atlas, exporting 3D model and so on.

When a new chunk is created it is activated automatically. Save project operation saves the content of all chunks. To save selected chunks as a separate project use Save Chunks command from the chunk context menu. To set another chunk as active 1. Right-click on the chunk title on the Workspace pane.

Select Set Active command from the context menu. To remove chunk 1. Select Remove Chunks command from the context menu. To rearrange the order of chunks in the Workspace pane simply drag and drop the chunks in the pane. Aligning chunks After the "partial" 3D models are built in several chunks they can be merged together.

Before merging chunks they need to be aligned. To align separate chunks 1. Select Align Chunks command from the Workflow menu. In the Align Chunks dialog box select chunks to be aligned, indicate reference chunk with a double- click. Set desired alignment options. To cancel processing click the Cancel button. Aligning chunks parameters The following parameters control the chunks alignment procedure and can be modified in the Align Chunks dialog box: Method Defines the chunks alignment method.

Point based method aligns chunks by matching photos across all the chunks. Camera based method is used to align chunks based on estimated camera locations. Corresponding cameras should have the same label.

Accuracy Point based alignment only Higher accuracy setting helps to obtain more accurate chunk alignment results. Lower accuracy setting can be used to get the rough chunk alignment in the shorter time. Point limit Point based alignment only The number indicates upper limit of feature points on every image to be taken into account during Point based chunks alignment. Fix scale Option is to be enabled in case the scales of the models in different chunks were set precisely and should be left unchanged during chunks alignment process.

A significant portion of this time is spent for matching of detected features across the photos. Image pair preselection option can speed up this process by selection of a subset of image pairs to be matched. Constrain features by mask Point based alignment only When this option is enabled, features detected in the masked image regions are discarded. For additional information on the usage of masks refer to the Using masks section. Merging chunks After alignment is complete the separate chunks can be merged into a single chunk.

To merge chunks 1. Select Merge Chunks command from the Workflow menu. In the Merge Chunks dialog box select chunks to be merged and the desired merging options. PhotoScan will merge the separate chunks into one.

The merged chunk will be displayed in the project content list on Workspace pane. The following parameters control the chunks merging procedure and can be modified in the Merge Chunks dialog box: Merge dense clouds Defines if dense clouds from the selected chunks are combined. Merge models Defines if models from the selected chunks are combined.

Chunks merging result i. Batch processing PhotoScan allows to perform general workflow operations with multiple chunks automatically. It is useful when dealing with a large number of chunks to be processed. Batch processing can be applied to all chunks in the Workspace, to unprocessed chunks only, or to the chunks selected by the user. Each operation chosen in the Batch processing dialog will be applied to every selected chunk before processing will move on to the next step.

Select Batch Process Click Add to add the desired processing stages. In the Add Job dialog select the kind of operation to be performed, the list of chunks it should be applied to, and desired processing parameters. Repeat the previous steps to add other processing steps as required. Arrange jobs by clicking Up and Down arrows at the right of the Batch Process Click OK button to start processing.

The progress dialog box will appear displaying the list and status of batch jobs and current operation progress. The list of tasks for batch processing can be exported to XML structured file using Save button in the Batch processing dialog and imported in a different project using Open button.

Graphical user interface Application window General view General view of application window. Model view Model view tab is used for displaying 3D data as well as for mesh and point cloud editing. The view of the model depends on the current processing stage and is also controlled by mode selection buttons on the PhotoScan toolbar. Along with the model the results of photo alignment can be displayed. These include sparse point cloud and camera positions visualised data. To enter the navigation mode click the Navigation toolbar button.

Photo view Photo view tab is used for displaying individual photos as well as masks on them. In the Photo view tab it is allowed to draw masks on the photos. Photo view is visible only if any photo is opened. To open a photo double-click on its name on the Workspace or Photos pane. Switching to Photo view mode changes the contents of the Toolbar, presenting related instruments and hiding irrelevant buttons. To select master channel 1.

Add multispectral images to the project using Add Photos Select Set Master Channel Display of images in PhotoScan window will be updated according to the master channel selection. Note Set Master Channel You can either indicate only one channel to be used as the basis for photogrammetric processing or leave the parameter value as Default for all three channels to be used in processing. When exporting in other formats, only master channel will be saved. Aligning photos Once photos are loaded into PhotoScan, they need to be aligned.

At this stage PhotoScan finds the camera position and orientation for each photo and builds a sparse point cloud model. To align a set of photos 1. Select Align Photos In the Align Photos dialog box select the desired alignment options. Click OK button when done. The progress dialog box will appear displaying the current processing status. To cancel processing click Cancel button.

You can inspect alignment results and remove incorrectly positioned photos, if any. To see the matches between any two photos use View Matches Incorrectly positioned photos can be realigned. To realign a subset of photos 1. Reset alignment for incorrectly positioned cameras using Reset Camera Alignment command from the photo context menu.

Set markers at least 4 per photo on these photos and indicate their projections on at least two photos from the already aligned subset. PhotoScan will consider these points to be true matches. For information on markers placement refer to the Setting coordinate system section. Select photos to be realigned and use Align Selected Cameras command from the photo context menu. When the alignment step is completed, the point cloud and estimated camera positions can be exported for processing with another software if needed.

Image quality Poor input, e. To help you to exclude poorly focused images from processing PhotoScan suggests automatic image quality estimation feature. Images with quality value of less than 0. To disable a photo use Disable button from the Photos pane toolbar. PhotoScan estimates image quality for each input image. The value of the parameter is calculated based on the sharpness level of the most focused part of the picture.

To estimate image quality 1. Switch to the detailed view in the Photos pane using on the Photos pane toolbar. Details command from the Change menu 2. Select all photos to be analyzed on the Photos pane. Right button click on the selected photo s and choose Estimate Image Quality command from the context menu.

Once the analysis procedure is over, a figure indicating estimated image quality value will be displayed in the Quality column on the Photos pane. Alignment parameters The following parameters control the photo alignment procedure and can be modified in the Align Photos dialog box: Accuracy Higher accuracy setting helps to obtain more accurate camera position estimates.

Lower accuracy setting can be used to get the rough camera positions in a shorter period of time. Pair preselection The alignment process of large photo sets can take a long time. A significant portion of this time period is spent on matching of detected features across the photos. Image pair preselection option may speed up this process due to selection of a subset of image pairs to be matched.

In the Generic preselection mode the overlapping pairs of photos are selected by matching photos using lower accuracy setting first. In the Reference preselection mode the overlapping pairs of photos are selected basing on the measured camera locations if present. For oblique imagery it is recommended to set Ground altitude value in the Settings dialog of the Reference pane to make the preselection procedure more efficient. Ground altitude information must be accompanied with yaw, pitch, roll data for cameras to be input in the Reference pane as well.

Additionally the following advanced parameters can be adjusted. Key point limit The number indicates upper limit of feature points on every image to be taken into account during current processing stage.

Using zero value allows PhotoScan to find as many key points as possible, but it may result in a big number of less reliable points.

Tie point limit The number indicates upper limit of matching points for every image. Using zero value doesn't apply any tie point filtering. Constrain features by mask When this option is enabled, features detected in the masked image regions are discarded. For additional information on the usage of masks please refer to the Using masks section. Note Tie point limit parameter allows to optimize performance for the task and does not generally effect the quality of the further model.

Recommended value is Too high tie-point limit value may cause some parts of the dense point cloud model to be missed. The reason is that PhotoScan generates depth maps only for pairs of photos for which number of matching points is above certain limit. As a results sparse point cloud will be thinned, yet the alignment will be kept unchanged.

Point cloud generation based on imported camera data PhotoScan supports import of external and internal camera orientation parameters. Thus, if for the project precise camera data is available, it is possible to load them into PhotoScan along with the photos, to be used as initial information for 3D reconstruction job.

To import external and internal camera parameters 1. Select Import Cameras command from the Tools menu. Select the format of the file to be imported. Browse to the file and click Open button. The data will be loaded into the software.

Camera calibration data can be inspected in the Camera Calibration dialog, Adjusted tab, available from Tools menu. If the input file contains some reference data camera position data in some coordinate system , the data will be shown on the Reference pane, View Estimated tab.

Once the data is loaded, PhotoScan will offer to build point cloud. This step involves feature points detection and matching procedures. As a result, a sparse point cloud - 3D representation of the tie-points data - will be generated.

Parameters controlling Build Point Cloud procedure are the same as the ones used at Align Photos step see above. Building dense point cloud PhotoScan allows to generate and visualize a dense point cloud model. Based on the estimated camera positions the program calculates depth information for each camera to be combined into a single dense point cloud.

PhotoScan tends to produce extra dense point clouds, which are of almost the same density, if not denser, as LIDAR point clouds. A dense point cloud can be edited and classified within PhotoScan environment or exported to an external tool for further analysis.

To build a dense point cloud 1. Check the reconstruction volume bounding box. To adjust the bounding box use the Resize Region and Rotate Region toolbar buttons. Rotate the bounding box and then drag corners of the box to the desired positions. Select the Build Dense Cloud In the Build Dense Cloud dialog box select the desired reconstruction parameters.

Reconstruction parameters Quality Specifies the desired reconstruction quality. Higher quality settings can be used to obtain more detailed and accurate geometry, but require longer time for processing. The only difference is that in this case Ultra High quality setting means processing of original photos, while each following step implies preprocessing image size downscaling by factor of 4 2 times by each side.

Depth Filtering modes At the stage of dense point cloud generation reconstruction PhotoScan calculates depth maps for every image. Due to some factors, like poor texture of some elements of the scene, noisy or badly focused images, there can be some outliers among the points. To sort out the outliers PhotoScan has several built-in filtering algorithms that answer the challenges of different projects.

If the area to be reconstructed does not contain meaningful small details, then it is reasonable to chose Aggressive depth filtering mode to sort out most of the outliers. Moderate depth filtering mode brings results that are in between the Mild and Aggressive approaches.

You can experiment with the setting in case you have doubts which mode to choose. Additionally depth filtering can be Disabled. But this option is not recommended as the resulting dense cloud could be extremely noisy. Building mesh To build a mesh 1. If the Height field reconstruction method is to be applied, it is important to control the position of the red side of the bounding box: it defines reconstruction plane.

In this case make sure that the bounding box is correctly oriented. Select the Build Mesh In the Build Mesh dialog box select the desired reconstruction parameters.

Reconstruction parameters PhotoScan supports several reconstruction methods and settings, which help to produce optimal reconstructions for a given data set. Surface type Arbitrary surface type can be used for modeling of any kind of object. It should be selected for closed objects, such as statues, buildings, etc.

It doesn't make any assumptions on the type of the object modeled, which comes at a cost of higher memory consumption. Height field surface type is optimized for modeling of planar surfaces, such as terrains or bas-reliefs. It should be selected for aerial photography processing as it requires lower amount of memory and allows for larger data sets processing. Source data Specifies the source for the mesh generation procedure.

Sparse cloud can be used for fast 3D model generation based solely on the sparse point cloud. Dense cloud setting will result in longer processing time but will generate high quality output based on the previously reconstructed dense point cloud.

Polygon count Specifies the maximum number of polygons in the final mesh. Suggested values High, Medium, Low are calculated based on the number of points in the previously generated dense point cloud: the They present optimal number of polygons for a mesh of a corresponding level of detail. It is still possible for a user to indicate the target number of polygons in the final mesh according to his choice.

It could be done through the Custom value of the Polygon count parameter. Please note that while too small number of polygons is likely to result in too rough mesh, too huge custom number over 10 million polygons is likely to cause model visualization problems in external software. Interpolation If interpolation mode is Disabled it leads to accurate reconstruction results since only areas corresponding to dense point cloud points are reconstructed.

Manual hole filling is usually required at the post processing step. With Enabled default interpolation mode PhotoScan will interpolate some surface areas within a circle of a certain radius around every dense cloud point. As a result some holes can be automatically covered. Yet some holes can still be present on the model and are to be filled at the post processing step. Enabled default setting is recommended for orthophoto generation. In Extrapolated mode the program generates holeless model with extrapolated geometry.

Large areas of extra geometry might be generated with this method, but they could be easily removed later using selection and cropping tools. Point classes Specifies the classes of the dense point cloud to be used for mesh generation. Preliminary dense cloud classification should be performed for this option of mesh generation to be active.

Note PhotoScan tends to produce 3D models with excessive geometry resolution, so it is recommended to perform mesh decimation after geometry computation. More information on mesh decimation and other 3D model geometry editing tools is given in the Editing model geometry section. Building model texture To generate 3D model texture 1. Select Build Texture Select the desired texture generation parameters in the Build Texture dialog box.

Texture mapping modes The texture mapping mode determines how the object texture will be packed in the texture atlas. Proper texture mapping mode selection helps to obtain optimal texture packing and, consequently, better visual quality of the final model. No assumptions regarding the type of the scene to be processed are made; program tries to create as uniform texture as possible. Adaptive orthophoto In the Adaptive orthophoto mapping mode the object surface is split into the flat part and vertical regions.

The flat part of the surface is textured using the orthographic projection, while vertical regions are textured separately to maintain accurate texture representation in such regions. When in the Adaptive orthophoto mapping mode, program tends to produce more compact texture representation for nearly planar scenes, while maintaining good texture quality for vertical surfaces, such as walls of the buildings.

Orthophoto In the Orthophoto mapping mode the whole object surface is textured in the orthographic projection. The Orthophoto mapping mode produces even more compact texture representation than the Adaptive orthophoto mode at the expense of texture quality in vertical regions.

Spherical Spherical mapping mode is appropriate only to a certain class of objects that have a ball-like form.

It allows for continuous texture atlas being exported for this type of objects, so that it is much easier to edit it later. When generating texture in Spherical mapping mode it is crucial to set the Bounding box properly. The whole model should be within the Bounding box. The red side of the Bounding box should be under the model; it defines the axis of the spherical projection.

The marks on the front side determine the 0 meridian. Single photo The Single photo mapping mode allows to generate texture from a single photo. The photo to be used for texturing can be selected from 'Texture from' list. Keep uv The Keep uv mapping mode generates texture atlas using current texture parametrization. It can be used to rebuild texture atlas using different resolution or to generate the atlas for the model parametrized in the external software.

Texture generation parameters The following parameters control various aspects of texture atlas generation: Texture from Single photo mapping mode only Specifies the photo to be used for texturing.

Available only in the Single photo mapping mode. Blending mode not used in Single photo mode Selects the way how pixel values from different photos will be combined in the final texture. Mosaic - gives more quality for orthophoto and texture atlas than Average mode, since it does not mix image details of overlapping photos but uses most appropriate photo i. Mosaic texture blending mode is especially useful for orthophoto generation based on approximate geometric model.

Average - uses the average value of all pixels from individual photos. Max Intensity - the photo which has maximum intensity of the corresponding pixel is selected. Min Intensity - the photo which has minimum intensity of the corresponding pixel is selected. Exporting texture to several files allows to archive greater resolution of the final model texture, while export of high resolution texture to a single file can fail due to RAM limitations.

Enable color correction The feature is useful for processing of data sets with extreme brightness variation. However, please note that color correction process takes up quite a long time, so it is recommended to enable the setting only for the data sets that proved to present results of poor quality.

To improve result texture quality it may be reasonable to exclude poorly focused images from processing at this step. PhotoScan suggests automatic image quality estimation feature. PhotoScan estimates image quality as a relative sharpness of the photo with respect to other images in the data set.

Saving intermediate results Certain stages of 3D model reconstruction can take a long time. The full chain of operations could easily last for hours when building a model from hundreds of photos. It is not always possible to finish all the operations in one run. PhotoScan allows to save intermediate results in a project file. PhotoScan project files may contain the following information: List of loaded photographs with reference paths to the image files.

Photo alignment data such as information on camera positions, sparse point cloud model and set of refined camera calibration parameters for each calibration group. Masks applied to the photos in project. Dense point cloud model with information on points classification.

Reconstructed 3D polygonal model with any changes made by user. This includes mesh and texture if it was built. List of added markers as well as of scale-bars and information on their positions.

Structure of the project, i. You can save the project at the end of any processing stage and return to it later. To restart work simply load the corresponding file into PhotoScan. Project files can also serve as backup files or be used to save different versions of the same model. Note that since PhotoScan tends to generate extra dense point clouds and highly detailed polygonal models, project saving procedure can take up quite a long time.

You can decrease compression level to speed up the saving process. However, please note that it will result in a larger project file. Compression level setting can be found on the Advanced tab of the Preferences dialog available from Tools menu. Project files use relative paths to reference original photos. Thus, when moving or copying the project file to another location do not forget to move or copy photographs with all the folder structure involved as well.

Otherwise, PhotoScan will fail to run any operation requiring source images, although the project file including the reconstructed model will be loaded up correctly.

Alternatively, you can enable Store absolute image paths option on the Advanced tab of the Preferences dialog available from Tools menu. Exporting results PhotoScan supports export of processing results in various representations: sparse and dense point clouds, camera calibration and camera orientation data, mesh, etc.

Point clouds and camera calibration data can be exported right after photo alignment is completed. All other export options are available after the 3D model is built. To align the model orientation with the default coordinate system use object button from the Toolbar.

Rotate In some cases editing model geometry in the external software may be required. PhotoScan supports model export for editing in external software and then allows to import it back, as it is described in the Editing model geometry section of the manual.

Main export commands are available from the File menu and the rest from the Export submenu of the Tools menu. Point cloud export To export sparse or dense point cloud 1. Select Export Points Browse the destination folder, choose the file type, and print in the file name. Click Save button. Specify the coordinate system and indicate export parameters applicable to the selected file type, including the dense cloud classes to be saved.

Click OK button to start export. Split in blocks option in the Export Points dialog can be useful for exporting large projects. It is available for referenced models only. You can indicate the size of the section in xy plane in meters for the point cloud to be divided into respective rectangular blocks.

The total volume of the 3D scene is limited with the Bounding Box. The whole volume will be split in equal blocks starting from the point with minimum x and y values. Note that empty blocks will not be saved. In some cases it may be reasonable to edit point cloud before exporting it. To read about point cloud editing refer to the Editing point cloud section of the manual.

Tie points data export To export matching points 1. Select Export Matches In the Export Matches dialog box set export parameters. Precision value sets the limit to the number of decimal digits in the tie points coordinates to be saved. Later on estimated camera data can be imported back to PhotoScan using Import Cameras command from the Tools menu to proceed with 3D model reconstruction procedure.

Camera calibration and orientation data export To export camera calibration and camera orientation data select Export Cameras Camera data export in Bundler file format would not save distortion coefficients k3, k4. PhotoScan is capable of panorama stitching for images taken from the same camera position - camera station. To indicate for the software that loaded images have been taken from one camera station one For information on camera group refer to Loading photos section.

To export panorama 1. Select Export - Export Panorama Select camera group which panorama should be previewed for. Choose panorama orientation in the file with the help of navigation buttons to the right of the preview window in the Export Panorama dialog. Set exporting parameters: select camera groups which panorama should be exported for and indicate export file name mask. Click OK button 6. Browse the destination folder and click Save button.

Select Export Model In the Export Model dialog specify the coordinate system and indicate export parameters applicable to the selected file type. If a model generated with PhotoScan is to be imported in a 3D editor program for inspection or further editing, it might be helpful to use Shift function while exporting the model.

It allows to set the value to be subtracted from the respective coordinate value for every vertex in the mesh. Essentially, this means translation of the model coordinate system origin, which may be useful since some 3D editors, for example, truncate the coordinates values up to 8 or so digits, while in some projects they are decimals that make sense with respect to model positioning task.

So it can be recommended to subtract a value equal to the whole part of a certain coordinate value see Reference pane, Camera coordinates values before exporting the model, thus providing for a reasonable scale for the model to be processed in a 3D editor program.

Phocus 2. Those consist out of two raster datasets,. Introduction This document details how to use the free software programme.

The estimated time to complete. GelAnalyzer User s manual Contents 1. Starting GelAnalyzer The main window Create a new analysis The image window Brian Caldwell, Ph. Introduction Digital photography has become a widely accepted alternative to conventional film photography for many applications ranging from.

Version 1. Added DNG image format support. Added support for changing chunks order in Workspace pane. Added drag-n-drop support for mesh import.

The material contained in this document is confidential and intended for use only by parties authorized by Gatewing. Visilter, S. Chapter 3 Starting ImageBrowser ImageBrowser Starting ImageBrowser 4 Importing Images to Your Computer FlexColor 3. About the Render Gallery All of your completed rendered images are available online from the Render Gallery page.

Images in the gallery are grouped in collections according to the source document RVT. Picture Manager Picture Manager allows you to easily edit and organize the pictures on your computer. Picture Manager is an application that was included with Microsoft Office suite for Windows and.

Digital Photogrammetric System Version 6. Purpose of the document General information The toolbar Adjustment batch mode Objects displaying. Operation is subject to the following two conditions: 1 This device may not. Operating system User Manual www. Appendix A. Then a propriety Data and. Contents 1. Introduction 1. Starting Publisher 2. Create a Poster Template 5.

Aligning your images and text 7. Apply a background Add text to your poster Add pictures to your poster Add graphs. F9 Integration Manager User Guide for use with QuickBooks This guide outlines the integration steps and processes supported for the purposes of financial reporting with F9 Professional and F9 Integration. Guide Contents Dear Milestone Customer, With the purchase of Milestone XProtect Central you have chosen a very powerful central monitoring solution, providing instant overview of any number of Milestone.

Start a new file in the Part. Interactive Voting System www. System Requirements Operation Flow

 


- Agisoft photoscan user manual professional edition version 1.3 free



 

Key point limit The number indicates upper limit of feature points on every image to be taken into account during current processing stage. Using zero value allows PhotoScan to find as many key points as possible, but it may result in a big number of less reliable points.

Tie point limit The number indicates upper limit of matching points for every image. Using zero value doesn't apply any tie point filtering. Constrain features by mask When this option is enabled, masked areas are excluded from feature detection procedure. For additional information on the usage of masks please refer to the Using masks section. Recommended value is Too high or too low tie point limit value may cause some parts of the dense point cloud model to be missed.

The reason is that PhotoScan generates depth maps only for pairs of photos for which number of matching points is above certain limit. As a results sparse point cloud will be thinned, yet the alignment will be kept unchanged. Point cloud generation based on imported camera data PhotoScan supports import of external and internal camera orientation parameters. Thus, if precise camera data is available for the project, it is possible to load them into PhotoScan along with the photos, to be used as initial information for 3D reconstruction job.

To import external and internal camera parameters 1. Select Import Cameras command from the Tools menu. Select the format of the file to be imported.

Browse to the file and click Open button. The data will be loaded into the software. Camera calibration data can be inspected in the Camera Calibration dialog, Adjusted tab, available from Tools menu. Once the data is loaded, PhotoScan will offer to build point cloud. This step involves feature points detection and matching procedures. As a result, a sparse point cloud - 3D representation of the tie-points data, will be generated.

Parameters controlling Build Point Cloud procedure are the same as the ones used at Align Photos step see above. Building dense point cloud PhotoScan allows to generate and visualize a dense point cloud model.

Based on the estimated camera positions the program calculates depth information for each camera to be combined into a single dense point cloud. PhotoScan tends to produce extra dense point clouds, which are of almost the same density, if not denser, as LIDAR point clouds. A dense point cloud can be edited within PhotoScan environment or exported to an external tool for further analysis. To build a dense point cloud 1. Check the reconstruction volume bounding box. To adjust the bounding box use the Resize Region and Rotate Region toolbar buttons.

Rotate the bounding box and then drag corners of the box to the desired positions. Select the Build Dense Cloud In the Build Dense Cloud dialog box select the desired reconstruction parameters. Reconstruction parameters Quality Specifies the desired reconstruction quality. Higher quality settings can be used to obtain more detailed and accurate geometry, but they require longer time for processing. Interpretation of the quality parameters here is similar to that of accuracy settings given in Photo Alignment section.

The only difference is that in this case Ultra High quality setting means processing of original photos, while each following step implies preliminary image size downscaling by factor of 4 2 times by each side. Depth Filtering modes At the stage of dense point cloud generation reconstruction PhotoScan calculates depth maps for every image. Due to some factors, like noisy or badly focused images, there can be some outliers among the points.

To sort out the outliers PhotoScan has several built-in filtering algorithms that answer the challenges of different projects. If there are important small details which are spatially distingueshed in the scene to be reconstructed, then it is recommended to set Mild depth filtering mode, for important features not to be sorted out as outliers. This value of the parameter may also be useful for aerial projects in case the area contains poorly textued roofs, for example.

If the area to be reconstructed does not contain meaningful small details, then it is reasonable to chose Aggressive depth filtering mode to sort out most of the outliers. This value of the parameter normally recommended for aerial data processing, however, mild filtering may be useful in some projects as well see poorly textured roofs comment in the mild parameter valur description above.

Moderate depth filtering mode brings results that are in between the Mild and Aggressive approaches.

You can experiment with the setting in case you have doubts which mode to choose. Additionally depth filtering can be Disabled. But this option is not recommended as the resulting dense cloud could be extremely noisy.

Building mesh To build a mesh 1. If the model has already been referenced, the bounding box will be properly positioned automatically. Otherwise, it is important to control its position manually. To adjust the bounding box manually, use the Resize Region and Rotate Region toolbar buttons. Rotate the bounding box and then drag corners of the box to the desired positions - only part of the scene inside the bounding box will be reconstructed.

If the Height field reconstruction method is to be applied, it is important to control the position of the red side of the bounding box: it defines reconstruction plane. In this case make sure that the bounding box is correctly oriented.

Select the Build Mesh In the Build Mesh dialog box select the desired reconstruction parameters. Reconstruction parameters PhotoScan supports several reconstruction methods and settings, which help to produce optimal reconstructions for a given data set. Surface type Arbitrary surface type can be used for modeling of any kind of object. It should be selected for closed objects, such as statues, buildings, etc.

It doesn't make any assumptions on the type of the object being modeled, which comes at a cost of higher memory consumption. Height field surface type is optimized for modeling of planar surfaces, such as terrains or basereliefs.

It should be selected for aerial photography processing as it requires lower amount of memory and allows for larger data sets processing. Source data Specifies the source for the mesh generation procedure.

Sparse cloud can be used for fast 3D model generation based solely on the sparse point cloud. Dense cloud setting will result in longer processing time but will generate high quality output based on the previously reconstructed dense point cloud. Polygon count Specifies the maximum number of polygons in the final mesh. They present optimal number of polygons for a mesh of a corresponding level of detail.

It is still possible for a user to indicate the target number of polygons in the final mesh according to their choice. It could be done through the Custom value of the Polygon count parameter. Please note that while too small number of polygons is likely to result in too rough mesh, too huge custom number over 10 million polygons is likely to cause model visualization problems in external software. Interpolation If interpolation mode is Disabled it leads to accurate reconstruction results since only areas corresponding to dense point cloud points are reconstructed.

Manual hole filling is usually required at the post processing step. With Enabled default interpolation mode PhotoScan will interpolate some surface areas within a circle of a certain radius around every dense cloud point. As a result some holes can be automatically covered. Yet some holes can still be present on the model and are to be filled at the post processing step. In Extrapolated mode the program generates holeless model with extrapolated geometry.

Large areas of extra geometry might be generated with this method, but they could be easily removed later using selection and cropping tools. More information on mesh decimation and other 3D model geometry editing tools is given in the Editing model geometry section.

Building model texture To generate 3D model texture 1. Select Build Texture Select the desired texture generation parameters in the Build Texture dialog box. Texture mapping modes The texture mapping mode determines how the object texture will be packed in the texture atlas. Proper texture mapping mode selection helps to obtain optimal texture packing and, consequently, better visual quality of the final model. Generic The default mode is the Generic mapping mode; it allows to parametrize texture atlas for arbitrary geometry.

No assumptions regarding the type of the scene to be processed are made; program tries to create as uniform texture as possible. Adaptive orthophoto In the Adaptive orthophoto mapping mode the object surface is split into the flat part and vertical regions. The flat part of the surface is textured using the orthographic projection, while vertical regions are textured separately to maintain accurate texture representation in such regions.

When in the Adaptive orthophoto mapping mode, program tends to produce more compact texture representation for nearly planar scenes, while maintaining good texture quality for vertical surfaces, such as walls of the buildings. Orthophoto In the Orthophoto mapping mode the whole object surface is textured in the orthographic projection.

The Orthophoto mapping mode produces even more compact texture representation than the Adaptive orthophoto mode at the expense of texture quality in vertical regions. Single photo The Single photo mapping mode allows to generate texture from a single photo. The photo to be used for texturing can be selected from 'Texture from' list. Keep uv The Keep uv mapping mode generates texture atlas using current texture parametrization.

It can be used to rebuild texture atlas using different resolution or to generate the atlas for the model parametrized in the external software. Texture generation parameters The following parameters control various aspects of texture atlas generation: Available only in the Single photo mapping mode.

Blending mode not used in Single photo mode Selects the way how pixel values from different photos will be combined in the final texture. Mosaic - implies two-step approach: it does blending of low frequency component for overlapping images to avoid seamline problem weighted average, weight being dependent on a number of parameters including proximity of the pixel in question to the center of the image , while high frequency component, that is in charge of picture details, is taken from a single image - the one that presents good resolution for the area of interest while the camera view is almost along the normal to the reconstructed surface in that point.

Average - uses the weighted average value of all pixels from individual photos, the weight being dependent on the same parameters that are considered for high frequence component in mosaic mode. Max Intensity - the photo which has maximum intensity of the corresponding pixel is selected.

Min Intensity - the photo which has minimum intensity of the corresponding pixel is selected. Disabled - the photo to take the color value for the pixel from is chosen like the one for the high frequency component in mosaic mode.

Exporting texture to several files allows to archive greater resolution of the final model texture, while export of high resolution texture to a single file can fail due to RAM limitations. Enable color correction The feature is useful for processing of data sets with extreme brightness variation. However, please note that color correction process takes up quite a long time, so it is recommended to enable the setting only for the data sets that proved to present results of poor quality.

Improving texture quality To improve resulting texture quality it may be reasonable to exclude poorly focused images from processing at this step. PhotoScan suggests automatic image quality estimation feature. PhotoScan estimates image quality as a relative sharpness of the photo with respect to other images in the data set.

Details command from the Change menu Saving intermediate results Certain stages of 3D model reconstruction can take a long time. The full chain of operations could eventually last for hours when building a model from hundreds of photos. It is not always possible to complete all the operations in one run. PhotoScan allows to save intermediate results in a project file. Photo alignment data such as information on camera positions, sparse point cloud model and set of refined camera calibration parameters for each calibration group.

Masks applied to the photos in project. Depth maps for cameras. Dense point cloud model. Reconstructed 3D polygonal model with any changes made by user. This includes mesh and texture if it was built. Structure of the project, i. Note that since PhotoScan tends to generate extra dense point clouds and highly detailed polygonal models, project saving procedure can take up quite a long time. You can decrease compression level to speed up the saving process. However, please note that it will result in a larger project file.

Compression level setting can be found on the Advanced tab of the Preferences dialog available from Tools menu. You can save the project at the end of any processing stage and return to it later. To restart work simply load the corresponding file into PhotoScan. Project files can also serve as backup files or be used to save different versions of the same model. Project files use relative paths to reference original photos.

Thus, when moving or copying the project file to another location do not forget to move or copy photographs with all the folder structure involved as well. Otherwise, PhotoScan will fail to run any operation requiring source images, although the project file including the reconstructed model will be loaded up correctly. Alternatively, you can enable Store absolute image paths option on the Advanced tab of the Preferences dialog available from Tools menu.

Exporting results PhotoScan supports export of processing results in various representations: sparse and dense point clouds, camera calibration and camera orientation data, mesh, etc. Point cloud and camera calibration data can be exported right after photo alignment is completed. All other export options are available after the corresponding processing step.

To align the model orientation with the default coordinate system use from the Toolbar. Rotate object button In some cases editing model geometry in the external software may be required.

PhotoScan supports model export for editing in external software and then allows to import it back as it is described in the Editing model geometry section of the manual. Main export commands are available from the File menu and the rest from the Export submenu of the Tools menu. Point cloud export To export sparse or dense point cloud 1. Select Export Points Browse the destination folder, choose the file type, and print in the file name.

Click Save button. Indicate export parameters applicable to the selected file type. Click OK button to start export. In some cases it may be reasonable to edit point cloud before exporting it.

To read about point cloud editing refer to the Editing point cloud section of the manual. Tie points data export To export matching points 1. Select Export Matches In the Export Matches dialog box set export parameters. Precision value sets the limit to the number of decimal digits in the tie points coordinates to be saved. Later on, estimated camera data can be imported back to PhotoScan using Import Cameras command from the Tools menu to proceed with 3D model reconstruction procedure.

Camera calibration and orientation data export To export camera calibration and camera orientation data select Export Cameras To indicate for the software that loaded images have been taken from one camera station, one should move those photos to a camera group and assign Camera Station type to it. For information on camera groups refer to Loading photos section.

To export panorama 1. Select Export - Export Panorama Select camera group which panorama should be previewed for. Choose panorama orientation in the file with the help of navigation buttons to the right of the preview window in the Export Panorama dialog.

Set exporting parameters: select camera groups which panorama should be exported for and indicate export file name mask. Click OK button 6. Browse the destination folder and click Save button. Additionally, you can set boundaries for the region of panorama to be exported using Setup boundaries section of the Export Panorama dialog. Text boxes in the first line X allow to indicate the angle in the horizontal plane and the second line Y serves for angle in the vertical plane limits. Image size option enables to control the size of the exporting file.

Select Export Model In the Export Model dialog indicate export parameters applicable to the selected file type. The texture file should be kept in the same directory as the main file describing the geometry. If the texture atlas was not built only the model geometry is exported. PhotoScan supports direct uploading of the models to Sketchfab resource. To publish your model online use Upload Model Extra products to export In addition to main targeted products PhotoScan allows to export several other processing results, like Undistort photos, i.

Depth map for any image Export Depth PhotoScan supports direct uploading of the models to Sketchfab resource and of the orthomosaics to MapBox platform. Improving camera alignment results Camera calibration Calibration groups While carrying out photo alignment PhotoScan estimates both internal and external camera orientation parameters, including nonlinear radial distortions.

For the estimation to be successful it is crucial to apply the estimation procedure separately to photos taken with different cameras. All the actions described below could and should be applied or not applied to each calibration group individually. Calibration groups can be rearranged manually. To create a new calibration group 1. Select Camera Calibration In the Camera Calibration dialog box, select photos to be arranged in a new group. In the right-click context menu choose Create Group command.

A new group will be created and depicted on the left-hand part of the Camera Calibration dialog box. To move photos from one group to another 1. In the Camera Calibration dialog box choose the source group on the left-hand part of the dialog. Select photos to be moved and drag them to the target group on the left-hand part of the Camera Calibration dialog box.

To place each photo into a separate group you can use Split Groups command available at the right button click on a calibration group name in the left-hand part of the Camera Calibration dialog Camera types PhotoScan supports two major types of camera: frame camera and fisheye camera. Camera type can be set in Camera Calibration dialog box available from Tools menu. Frame camera. If the source data within a calibration group was shot with a frame camera, for successful estimation of camera orientation parameters the information on approximate focal length pix is required.

Obviously, to calculate focal length value in pixel it is enough to know focal length in mm along with the sensor pixel size in mm. Normally this data is extracted automatically from the EXIF metadata. Frame camera with Fisheye lens. If extra wide lenses were used to get the source data, standard PhotoScan camera model will not allow to estimate camera parameters successfully.

Fisheye camera type setting will initialize implementation of a different camera model to fit ultra-wide lens distortions. Those consist out of two raster datasets,. Phocus 2. Version 1. Added DNG image format support. Added support for changing chunks order in Workspace pane.

Added drag-n-drop support for mesh import. Introduction This document details how to use the free software programme. GelAnalyzer User s manual Contents 1. Starting GelAnalyzer The main window Create a new analysis The image window The estimated time to complete.

Chapter 3 Starting ImageBrowser ImageBrowser Starting ImageBrowser 4 Importing Images to Your Computer About the Render Gallery All of your completed rendered images are available online from the Render Gallery page. Images in the gallery are grouped in collections according to the source document RVT. The material contained in this document is confidential and intended for use only by parties authorized by Gatewing. Visilter, S. Picture Manager Picture Manager allows you to easily edit and organize the pictures on your computer.

Picture Manager is an application that was included with Microsoft Office suite for Windows and. Brian Caldwell, Ph.

Introduction Digital photography has become a widely accepted alternative to conventional film photography for many applications ranging from.

Operation is subject to the following two conditions: 1 This device may not. Contents 1. Introduction 1. Starting Publisher 2. Create a Poster Template 5. Aligning your images and text 7.

Apply a background Add text to your poster Add pictures to your poster Add graphs. Using Microsoft Word Many Word documents will require elements that were created in programs other than Word, such as the picture to the right. Nontext elements in a document are referred to as Objects. Contents Introduction to the TI Connect 4. Page Properties 1. Type the name of the Title of the page, for example Template. And then click OK.

User Manual www. When you specify a high resolution,. Avidemux, to exerpt the video clip, read the video properties, and save. Introduction The Fixplot program is a component program of Eyenal that allows the user to plot eye position data collected with.

FlexColor 3. A technical overview of the Fuel3D system. Contents Introduction 3 How does Fuel3D actually work? Appendix A. Microsoft PowerPoint is used for creating slide presentations to be used on a PC or even printed out as overheads and handouts. You can combine text, graphics, even audio and video to create exciting and.

All rights reserved. Information in this document is subject to change without notice. Protocol for Microscope Calibration A properly calibrated system is essential for successful and efficient software use.

The following are step by step instructions on how to calibrate the hardware using. Operating system Information might have been changed or added since this document. Chapter 2 Tutorial Tutorial Introduction This tutorial is designed to introduce you to some of Surfer's basic features. After you have completed the tutorial, you should be able to begin creating your.

Then a propriety Data and. Under no circumstances. Please refer to the appropriate. System Requirements Operation Flow The main imovie window is divided into six major parts. Project Drag clips to the project area to create a timeline 2. Preview Window Displays a preview of your video 3. Toolbar Contains a variety of. The information in this publication is covered under Legal Notices for this product.

Stathopoulou a, A. Valanis a, J. Lerma b, A. Guide Contents Dear Milestone Customer, With the purchase of Milestone XProtect Central you have chosen a very powerful central monitoring solution, providing instant overview of any number of Milestone. It is fast, powerful and easy-to-use for every expert and beginners.

Starting DWGSee. These instructions are for an Epson. M Image Protection Mechanism setting. Introduction VrayPattern is a plug-in for 3dsmax and V-Ray VRayPattern allows to multiply geometry periodically without extra memory consumption. As a sample you can use any tiled geometry or those with.

How to resize, rotate, and crop images You will frequently want to resize and crop an image after opening it in Photoshop from a digital camera or scanner. Cropping means cutting some parts of the image. Bitrix Site Manager 4. F9 Integration Manager User Guide for use with QuickBooks This guide outlines the integration steps and processes supported for the purposes of financial reporting with F9 Professional and F9 Integration. On the contrary, the set of camera positions is required for further 3D model construction by PhotoScan.

The next stage is building dense point cloud. Based on the estimated camera positions and pictures themselves a dense point cloud is built by PhotoScan. Dense point cloud may be edited and classified prior to export or proceeding to 3D mesh model generation. Another stage is building mesh. PhotoScan reconstructs a 3D polygonal mesh representing the object surface based on the dense point cloud. Additionally there is a Point Cloud based method for fast geometry generation based on the sparse point cloud alone.

Generally there are two algorithmic methods available in PhotoScan that can be applied to 3D mesh generation: Height Field - for planar type surfaces, Arbitrary - for any kind of object.

Having built the mesh, it may be necessary to edit it. Some corrections, such as mesh decimation, removal of detached components, closing of holes in the mesh, etc. For more complex editing you have to engage external 3D editor tools.

PhotoScan allows to export mesh, edit it by another software and import it back. After geometry i. Several texturing modes are available in PhotoScan, they are described in the corresponding section of this manual. About the manual Basically, the sequence of actions described above covers most of the data processing needs. All these operations are carried out automatically according to the parameters set by user.

Instructions on how to get through these operations and descriptions of the parameters controlling each step are explained in the corresponding sections of the Chapter 3, General workflow. In some cases, however, additional actions may be required to get the desired results. In some capturing scenarios masking of certain regions of the photos may be required to exclude them from the calculations.

Application of masks in PhotoScan processing workflow as well as editing options available are v. Camera calibration issues are discussed in Chapter 4, Referencing and measurements, that also describes functionality to reference the results and carry out measurements on the model.

While Chapter 6, Automation describes opportunities to save up on manual intervention to the processing workflow, Chapter 7, Network processing presents guidelines on how to organize distributed processing of the imagery data on several nodes.

It can take up quite a long time to reconstruct a 3D model. PhotoScan allows to export obtained results and save intermediate data in a form of project files at any stage of the process. If you are not familiar with the concept of projects, its brief description is given at the end of the Chapter 3, General workflow.

In the manual you can also find instructions on the PhotoScan installation procedure and basic rules for taking "good" photographs, i. For the information refer to Chapter 2, Capturing photos and Chapter 1, Installation. NVidia GeForce 8xxx series and later. PhotoScan is likely to be able to utilize processing power of any OpenCL enabled device during Dense Point Cloud generation stage, provided that OpenCL drivers for the device are properly installed.

However, because of the large number of various combinations of video chips, driver versions and operating systems, Agisoft is unable to test and guarantee PhotoScan's compatibility with every device and on every platform.

The table below lists currently supported devices on Windows platform only. We will pay particular attention to possible problems with PhotoScan running on these devices. Table 1. Using OpenCL acceleration with mobile or integrated graphics video chips is not recommended because of the low performance of such GPUs.

Start PhotoScan by running photoscan. Restrictions of the Demo mode Once PhotoScan is downloaded and installed on your computer you can run it either in the Demo mode or in the full function mode.

On every start until you enter a serial number it will show a registration box offering two options: 1 use PhotoScan in the Demo mode or 2 enter a serial number to confirm the purchase. The first choice is set by default, so if you are still exploring PhotoScan click the Continue button and PhotoScan will start in the Demo mode. The employment of PhotoScan in the Demo mode is not time limited. Several functions, however, are not available in the Demo mode.

These functions are the following: saving the project; 2. To use PhotoScan in the full function mode you have to purchase it. On purchasing you will get the serial number to enter into the registration box on starting PhotoScan. Once the serial number is entered the registration box will not appear again and you will get full access to all functions of the program.

Capturing photos Before loading your photographs into PhotoScan you need to take them and select those suitable for 3D model reconstruction. Photographs can be taken by any digital camera both metric and non-metric , as long as you follow some specific capturing guidelines. This section explains general principles of taking and selecting pictures that provide the most appropriate data for 3D model generation. Make sure you have studied the following rules and read the list of restrictions before you get out for shooting photographs.

Equipment Use a digital camera with reasonably high resolution 5 MPix or more. Avoid ultra-wide angle and fish-eye lenses. The best choice is 50 mm focal length 35 mm film equivalent lenses. It is recommended to use focal length from 20 to 80 mm interval in 35mm equivalent. If a data set was captured with fish-eye lens, appropriate camera sensor type should be selected in PhotoScan Camera Calibration settings prior to processing.

Fixed lenses are preferred. If zoom lenses are used - focal length should be set either to maximal or to minimal value during the entire shooting session for more stable results.

Take images at maximal possible resolution. ISO should be set to the lowest value, otherwise high ISO values will induce additional noise to images. Aperture value should be high enough to result in sufficient focal depth: it is important to capture sharp, not blurred photos.

Shutter speed should not be too slow, otherwise blur can occur due to slight movements. If still have to, shoot shiny objects under a cloudy sky. Avoid unwanted foregrounds. Avoid moving objects within the scene to be reconstructed. Avoid absolutely flat objects or scenes. Image preprocessing PhotoScan operates with the original images. So do not crop or geometrically transform, i. Number of photos: more than required is better than not enough. Number of "blind-zones" should be minimized since PhotoScan is able to reconstruct only geometry visible from at least two cameras.

Each photo should effectively use the frame size: object of interest should take up the maximum area. In some cases portrait camera orientation should be used. Do not try to place full object in the image frame, if some parts are missing it is not a problem providing that these parts appear on other images. Good lighting is required to achieve better quality of the results, yet blinks should be avoided. It is recommended to remove sources of light from camera fields of view.

Avoid using flash. If you are planning to carry out any measurements based on the reconstructed model, do not forget to locate at least two markers with a known distance between them on the object. Alternatively, you could place a ruler within the shooting area. In case of aerial photography and demand to fulfill georeferencing task, even spread of ground control points GCPs at least 10 across the area to be reconstructed is required to achieve results of highest quality, both in terms of the geometrical precision and georeferencing accuracy.

The following figures represent advice on appropriate capturing scenarios: Facade Incorrect Facade Correct 5. A short list of typical reasons for photographs unsuitability is given below. Modifications of photographs PhotoScan can process only unmodified photos as they were taken by a digital photo camera. Processing the photos which were manually cropped or geometrically warped is likely to fail or to produce highly inaccurate results.

Photometric modifications do not affect reconstruction results. In this case PhotoScan assumes that focal length in 35 mm equivalent equals to 50 mm and tries to align the photos in accordance with this assumption. If the correct focal length value differs significantly from 50 mm, the alignment can give incorrect results or even fail. In such cases it is required to specify initial camera calibration manually. The details of necessary EXIF tags and instructions for manual setting of the calibration parameters are given in the Camera calibration section.

Otherwise it is most unlikely that processing results will be accurate. Fisheye and ultra-wide angle lenses are poorly modeled by the common distortion model implemented in PhotoScan software, so it is required to choose proper camera type in Camera Calibration dialog prior to processing. General workflow Processing of images with PhotoScan includes the following main steps: loading photos into PhotoScan; inspecting loaded images, removing unnecessary images; aligning photos; building dense point cloud; building mesh 3D polygonal model ; generating texture; exporting results.

If you are using PhotoScan in the full function not the Demo mode, intermediate results of the image processing can be saved at any stage in the form of project files and can be used later. The concept of projects and project files is briefly explained in the Saving intermediate results section. The list above represents all the necessary steps involved in the construction of a textured 3D model from your photos.

Some additional tools, which you may find to be useful, are described in the successive chapters. Preferences settings Before starting a project with PhotoScan it is recommended to adjust the program settings for your needs. In Preferences dialog General Tab available through the Tools menu you can indicate the path to the PhotoScan log file to be shared with the Agisoft support team in case you face any problems during the processing. Here you can also change GUI language to the one that is most convenient for you.

PhotoScan exploits GPU processing power that speeds up the process significantly. If you have decided to switch on GPUs for photogrammetric data processing with PhotoScan, it is recommended to free one physical CPU core per each active GPU for overall control and resource managing tasks.

Loading photos Before starting any operation it is necessary to point out what photos will be used as a source for 3D reconstruction. In fact, photographs themselves are not loaded into PhotoScan until they are needed. So, when you "load photos" you only indicate photographs that will be used for further processing. To load a set of photos 1. Select Add Photos Add Photos toolbar button on 2.

In the Add Photos dialog box browse to the folder containing the images and select files to be processed. Then click Open button. Selected photos will appear on the Workspace pane. Photos in any other format will not be shown in the Add Photos dialog box. To work with such photos you will need to convert them in one of the supported formats. If you have loaded some unwanted photos, you can easily remove them at any moment.

To remove unwanted photos 1. On the Workspace pane select the photos to be removed. Right-click on the selected photos and choose Remove Items command from the opened context Camera groups menu, or click Remove Items toolbar button on the Workspace pane. The selected photos will be removed from the working set. If all the photos or a subset of photos were captured from one camera position - camera station, for PhotoScan to process them correctly it is obligatory to move those photos to a camera group and mark the group as Camera Station.

It is important that for all the photos in a Camera Station group distances between camera centers were negligibly small compared to the camera-object minimal distance. However, it is possible to export panoramic picture for the data captured from only one camera station. Refer to Exporting results section for guidance on panorama export. To move photos to a camera group 1. On the Workspace pane or Photos pane select the photos to be moved. Right-click on the selected photos and choose Move Cameras - New Camera Group command from the opened context menu.

A new group will be added to the active chunk structure and selected photos will be moved to that group. To mark group as camera station right click on the camera group name and select Set Group Type command from the context menu.

Inspecting loaded photos Loaded photos are displayed on the Workspace pane along with flags reflecting their status. The following flags can appear next to the photo name: NC Not calibrated Notifies that the EXIF data available is not sufficient to estimate the camera focal length. In this case PhotoScan assumes that the corresponding photo was taken using 50mm lens 35mm film equivalent.

More details on manual camera calibration can be found in the Camera calibration section. NA Not aligned Notifies that external camera orientation parameters were not estimated for the current photo yet.

Images loaded to PhotoScan will not be aligned until you perform the next step - photos alignment. Notifies that Camera Station type was assigned to the group.

Multispectral imagery PhotoScan supports processing of multispectral images saved as multichannel single page TIFF files. The main processing stages for multispectral images are performed based on the master channel, which can be selected by the user. During orthophoto export, all spectral bands are processed together to form a multispectral orthophoto with the same bands as in source images.

The overall procedure for multispectral imagery processing does not differ from the usual procedure for normal photos, except the additional master channel selection step performed after adding images to the project. For the best results it is recommended to select the spectral band which is sharp and as much detailed as possible. To select master channel 1. Add multispectral images to the project using Add Photos Select Set Master Channel Display of images in PhotoScan window will be updated according to the master channel selection.

Note Set Master Channel You can either indicate only one channel to be used as the basis for photogrammetric processing or leave the parameter value as Default for all three channels to be used in processing. When exporting in other formats, only master channel will be saved. Aligning photos Once photos are loaded into PhotoScan, they need to be aligned. At this stage PhotoScan finds the camera position and orientation for each photo and builds a sparse point cloud model.

To align a set of photos 1. Select Align Photos In the Align Photos dialog box select the desired alignment options. Click OK button when done. The progress dialog box will appear displaying the current processing status. To cancel processing click Cancel button. You can inspect alignment results and remove incorrectly positioned photos, if any.

To see the matches between any two photos use View Matches Incorrectly positioned photos can be realigned. To realign a subset of photos 1.

Reset alignment for incorrectly positioned cameras using Reset Camera Alignment command from the photo context menu. Set markers at least 4 per photo on these photos and indicate their projections on at least two photos from the already aligned subset.

PhotoScan will consider these points to be true matches. For information on markers placement refer to the Setting coordinate system section. Select photos to be realigned and use Align Selected Cameras command from the photo context menu. When the alignment step is completed, the point cloud and estimated camera positions can be exported for processing with another software if needed. Image quality Poor input, e. To help you to exclude poorly focused images from processing PhotoScan suggests automatic image quality estimation feature.

Images with quality value of less than 0. To disable a photo use Disable button from the Photos pane toolbar. PhotoScan estimates image quality for each input image. The value of the parameter is calculated based on the sharpness level of the most focused part of the picture.

To estimate image quality 1. Switch to the detailed view in the Photos pane using on the Photos pane toolbar. Details command from the Change menu 2. Select all photos to be analyzed on the Photos pane. Right button click on the selected photo s and choose Estimate Image Quality command from the context menu. Once the analysis procedure is over, a figure indicating estimated image quality value will be displayed in the Quality column on the Photos pane.

Alignment parameters The following parameters control the photo alignment procedure and can be modified in the Align Photos dialog box: Accuracy Higher accuracy setting helps to obtain more accurate camera position estimates. Lower accuracy setting can be used to get the rough camera positions in a shorter period of time. Pair preselection The alignment process of large photo sets can take a long time. A significant portion of this time period is spent on matching of detected features across the photos.

Image pair preselection option may speed up this process due to selection of a subset of image pairs to be matched. In the Generic preselection mode the overlapping pairs of photos are selected by matching photos using lower accuracy setting first.

In the Reference preselection mode the overlapping pairs of photos are selected basing on the measured camera locations if present.

For oblique imagery it is recommended to set Ground altitude value in the Settings dialog of the Reference pane to make the preselection procedure more efficient. Ground altitude information must be accompanied with yaw, pitch, roll data for cameras to be input in the Reference pane as well. Additionally the following advanced parameters can be adjusted. Key point limit The number indicates upper limit of feature points on every image to be taken into account during current processing stage.

Using zero value allows PhotoScan to find as many key points as possible, but it may result in a big number of less reliable points. Tie point limit The number indicates upper limit of matching points for every image. Using zero value doesn't apply any tie point filtering. Constrain features by mask When this option is enabled, features detected in the masked image regions are discarded.

For additional information on the usage of masks please refer to the Using masks section. Note Tie point limit parameter allows to optimize performance for the task and does not generally effect the quality of the further model. Recommended value is Too high tie-point limit value may cause some parts of the dense point cloud model to be missed.

The reason is that PhotoScan generates depth maps only for pairs of photos for which number of matching points is above certain limit. As a results sparse point cloud will be thinned, yet the alignment will be kept unchanged. Point cloud generation based on imported camera data PhotoScan supports import of external and internal camera orientation parameters.

Thus, if for the project precise camera data is available, it is possible to load them into PhotoScan along with the photos, to be used as initial information for 3D reconstruction job. To import external and internal camera parameters 1. Select Import Cameras command from the Tools menu.

Select the format of the file to be imported. Browse to the file and click Open button. The data will be loaded into the software. Camera calibration data can be inspected in the Camera Calibration dialog, Adjusted tab, available from Tools menu. If the input file contains some reference data camera position data in some coordinate system , the data will be shown on the Reference pane, View Estimated tab.

Once the data is loaded, PhotoScan will offer to build point cloud. This step involves feature points detection and matching procedures. As a result, a sparse point cloud - 3D representation of the tie-points data - will be generated. Parameters controlling Build Point Cloud procedure are the same as the ones used at Align Photos step see above.

Building dense point cloud PhotoScan allows to generate and visualize a dense point cloud model. Based on the estimated camera positions the program calculates depth information for each camera to be combined into a single dense point cloud.

PhotoScan tends to produce extra dense point clouds, which are of almost the same density, if not denser, as LIDAR point clouds. A dense point cloud can be edited and classified within PhotoScan environment or exported to an external tool for further analysis.

To build a dense point cloud 1. Check the reconstruction volume bounding box. To adjust the bounding box use the Resize Region and Rotate Region toolbar buttons. Rotate the bounding box and then drag corners of the box to the desired positions.

Select the Build Dense Cloud In the Build Dense Cloud dialog box select the desired reconstruction parameters. Reconstruction parameters Quality Specifies the desired reconstruction quality. Higher quality settings can be used to obtain more detailed and accurate geometry, but require longer time for processing.

The only difference is that in this case Ultra High quality setting means processing of original photos, while each following step implies preprocessing image size downscaling by factor of 4 2 times by each side.

Depth Filtering modes At the stage of dense point cloud generation reconstruction PhotoScan calculates depth maps for every image. Due to some factors, like poor texture of some elements of the scene, noisy or badly focused images, there can be some outliers among the points. To sort out the outliers PhotoScan has several built-in filtering algorithms that answer the challenges of different projects.

If the area to be reconstructed does not contain meaningful small details, then it is reasonable to chose Aggressive depth filtering mode to sort out most of the outliers. Moderate depth filtering mode brings results that are in between the Mild and Aggressive approaches. You can experiment with the setting in case you have doubts which mode to choose. Additionally depth filtering can be Disabled. But this option is not recommended as the resulting dense cloud could be extremely noisy.

Building mesh To build a mesh 1. If the Height field reconstruction method is to be applied, it is important to control the position of the red side of the bounding box: it defines reconstruction plane. In this case make sure that the bounding box is correctly oriented. Select the Build Mesh In the Build Mesh dialog box select the desired reconstruction parameters.

Reconstruction parameters PhotoScan supports several reconstruction methods and settings, which help to produce optimal reconstructions for a given data set. Surface type Arbitrary surface type can be used for modeling of any kind of object. It should be selected for closed objects, such as statues, buildings, etc.

It doesn't make any assumptions on the type of the object modeled, which comes at a cost of higher memory consumption. Height field surface type is optimized for modeling of planar surfaces, such as terrains or bas-reliefs. It should be selected for aerial photography processing as it requires lower amount of memory and allows for larger data sets processing. Source data Specifies the source for the mesh generation procedure. Sparse cloud can be used for fast 3D model generation based solely on the sparse point cloud.

Dense cloud setting will result in longer processing time but will generate high quality output based on the previously reconstructed dense point cloud. Polygon count Specifies the maximum number of polygons in the final mesh. Suggested values High, Medium, Low are calculated based on the number of points in the previously generated dense point cloud: the They present optimal number of polygons for a mesh of a corresponding level of detail.

It is still possible for a user to indicate the target number of polygons in the final mesh according to his choice. It could be done through the Custom value of the Polygon count parameter.

Please note that while too small number of polygons is likely to result in too rough mesh, too huge custom number over 10 million polygons is likely to cause model visualization problems in external software. Interpolation If interpolation mode is Disabled it leads to accurate reconstruction results since only areas corresponding to dense point cloud points are reconstructed.

Manual hole filling is usually required at the post processing step. With Enabled default interpolation mode PhotoScan will interpolate some surface areas within a circle of a certain radius around every dense cloud point. As a result some holes can be automatically covered. Yet some holes can still be present on the model and are to be filled at the post processing step.

Enabled default setting is recommended for orthophoto generation. In Extrapolated mode the program generates holeless model with extrapolated geometry. Large areas of extra geometry might be generated with this method, but they could be easily removed later using selection and cropping tools.

Point classes Specifies the classes of the dense point cloud to be used for mesh generation. Preliminary dense cloud classification should be performed for this option of mesh generation to be active. Note PhotoScan tends to produce 3D models with excessive geometry resolution, so it is recommended to perform mesh decimation after geometry computation.

More information on mesh decimation and other 3D model geometry editing tools is given in the Editing model geometry section.

Building model texture To generate 3D model texture 1. Select Build Texture Select the desired texture generation parameters in the Build Texture dialog box. Texture mapping modes The texture mapping mode determines how the object texture will be packed in the texture atlas.

Proper texture mapping mode selection helps to obtain optimal texture packing and, consequently, better visual quality of the final model. No assumptions regarding the type of the scene to be processed are made; program tries to create as uniform texture as possible. Adaptive orthophoto In the Adaptive orthophoto mapping mode the object surface is split into the flat part and vertical regions. The flat part of the surface is textured using the orthographic projection, while vertical regions are textured separately to maintain accurate texture representation in such regions.

When in the Adaptive orthophoto mapping mode, program tends to produce more compact texture representation for nearly planar scenes, while maintaining good texture quality for vertical surfaces, such as walls of the buildings. Orthophoto In the Orthophoto mapping mode the whole object surface is textured in the orthographic projection. The Orthophoto mapping mode produces even more compact texture representation than the Adaptive orthophoto mode at the expense of texture quality in vertical regions.

Spherical Spherical mapping mode is appropriate only to a certain class of objects that have a ball-like form. It allows for continuous texture atlas being exported for this type of objects, so that it is much easier to edit it later. When generating texture in Spherical mapping mode it is crucial to set the Bounding box properly.

The whole model should be within the Bounding box. The red side of the Bounding box should be under the model; it defines the axis of the spherical projection. The marks on the front side determine the 0 meridian. Single photo The Single photo mapping mode allows to generate texture from a single photo.

   


Comments

Popular posts from this blog

15 Best Online Free Resources For Mastering AutoCAD

- Free online microsoft office word 2007 free

- Adobe Photoshop CS Portable(32 Bits wenjarei ⚡ - Wakelet