Many tropical countries suffer from persistent cloud cover inhibiting spatially consistent reporting of deforestation and forest degradation for REDD+. Data gaps remain even when compositing Landsat-like optical satellite imagery over one or two years. Instead, medium resolution SAR is capable of providing reliable deforestation information but shows limited capacity to identify forest degradation. This paper describes an innovative approach for feature fusion of multi-temporal and medium-resolution SAR and optical sub-pixel fraction information. After independently processing SAR and optical input data streams the extracted SAR and optical sub-pixel fraction features are fused using a decision tree classifier. ALOS PALSAR Fine Bean Dual and Landsat imagery of 2007 and 2010 acquired over the main mining district in central Guyana have been used for a proof-of-concept demonstration observing overall accuracies of 88% and 89.3% formapping forest land cover and detecting deforestation and forest degradation, respectively. Deforestation and degradation rates of 0.1% and 0.08% are reported for the observation period. Data gaps due to mainly clouds and Landsat ETM+ SLC-off that remained after compositing a set of single-period Landsat scenes, but also due to SAR layover and shadow could be reduced from 7.9% to negligible 0.01% while maintaining the desired thematic detail of detecting deforestation and degradation. The paper demonstrates the increase of both spatial completeness and thematic detail when applying the methodology, compared with potential Landsat-only or PALSAR-only approaches for a heavy cloud contaminated tropical environment. It indicates the potential for providing the required accuracy of activity data for REDD+ MRV.